00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-v23.11" build number 925 00:00:00.001 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3586 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.001 Started by timer 00:00:00.023 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.025 The recommended git tool is: git 00:00:00.025 using credential 00000000-0000-0000-0000-000000000002 00:00:00.027 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.041 Fetching changes from the remote Git repository 00:00:00.046 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.063 Using shallow fetch with depth 1 00:00:00.063 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.063 > git --version # timeout=10 00:00:00.086 > git --version # 'git version 2.39.2' 00:00:00.086 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.116 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.116 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.940 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.951 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.962 Checking out Revision 44e7d6069a399ee2647233b387d68a938882e7b7 (FETCH_HEAD) 00:00:02.962 > git config core.sparsecheckout # timeout=10 00:00:02.972 > git read-tree -mu HEAD # timeout=10 00:00:02.987 > git checkout -f 44e7d6069a399ee2647233b387d68a938882e7b7 # timeout=5 00:00:03.004 Commit message: "scripts/bmc: Rework Get NIC Info cmd parser" 00:00:03.004 > git rev-list --no-walk 44e7d6069a399ee2647233b387d68a938882e7b7 # timeout=10 00:00:03.173 [Pipeline] Start of Pipeline 00:00:03.188 [Pipeline] library 00:00:03.190 Loading library shm_lib@master 00:00:03.190 Library shm_lib@master is cached. Copying from home. 00:00:03.207 [Pipeline] node 00:00:03.227 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:03.229 [Pipeline] { 00:00:03.237 [Pipeline] catchError 00:00:03.238 [Pipeline] { 00:00:03.250 [Pipeline] wrap 00:00:03.258 [Pipeline] { 00:00:03.266 [Pipeline] stage 00:00:03.268 [Pipeline] { (Prologue) 00:00:03.285 [Pipeline] echo 00:00:03.287 Node: VM-host-WFP7 00:00:03.292 [Pipeline] cleanWs 00:00:03.300 [WS-CLEANUP] Deleting project workspace... 00:00:03.300 [WS-CLEANUP] Deferred wipeout is used... 00:00:03.306 [WS-CLEANUP] done 00:00:03.495 [Pipeline] setCustomBuildProperty 00:00:03.562 [Pipeline] httpRequest 00:00:04.215 [Pipeline] echo 00:00:04.216 Sorcerer 10.211.164.101 is alive 00:00:04.226 [Pipeline] retry 00:00:04.227 [Pipeline] { 00:00:04.240 [Pipeline] httpRequest 00:00:04.245 HttpMethod: GET 00:00:04.246 URL: http://10.211.164.101/packages/jbp_44e7d6069a399ee2647233b387d68a938882e7b7.tar.gz 00:00:04.246 Sending request to url: http://10.211.164.101/packages/jbp_44e7d6069a399ee2647233b387d68a938882e7b7.tar.gz 00:00:04.247 Response Code: HTTP/1.1 200 OK 00:00:04.247 Success: Status code 200 is in the accepted range: 200,404 00:00:04.248 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_44e7d6069a399ee2647233b387d68a938882e7b7.tar.gz 00:00:04.524 [Pipeline] } 00:00:04.538 [Pipeline] // retry 00:00:04.543 [Pipeline] sh 00:00:04.823 + tar --no-same-owner -xf jbp_44e7d6069a399ee2647233b387d68a938882e7b7.tar.gz 00:00:04.838 [Pipeline] httpRequest 00:00:05.178 [Pipeline] echo 00:00:05.180 Sorcerer 10.211.164.101 is alive 00:00:05.189 [Pipeline] retry 00:00:05.191 [Pipeline] { 00:00:05.205 [Pipeline] httpRequest 00:00:05.210 HttpMethod: GET 00:00:05.211 URL: http://10.211.164.101/packages/spdk_12fc2abf1e54ef44d6ae9091ab879722d4e15e60.tar.gz 00:00:05.211 Sending request to url: http://10.211.164.101/packages/spdk_12fc2abf1e54ef44d6ae9091ab879722d4e15e60.tar.gz 00:00:05.212 Response Code: HTTP/1.1 200 OK 00:00:05.213 Success: Status code 200 is in the accepted range: 200,404 00:00:05.213 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_12fc2abf1e54ef44d6ae9091ab879722d4e15e60.tar.gz 00:00:39.477 [Pipeline] } 00:00:39.496 [Pipeline] // retry 00:00:39.504 [Pipeline] sh 00:00:39.792 + tar --no-same-owner -xf spdk_12fc2abf1e54ef44d6ae9091ab879722d4e15e60.tar.gz 00:00:42.344 [Pipeline] sh 00:00:42.629 + git -C spdk log --oneline -n5 00:00:42.629 12fc2abf1 test: Remove autopackage.sh 00:00:42.629 83ba90867 fio/bdev: fix typo in README 00:00:42.630 45379ed84 module/compress: Cleanup vol data, when claim fails 00:00:42.630 0afe95a3a bdev/nvme: use bdev_nvme linker script 00:00:42.630 1cbacb58f test/nvmf: Clarify comment about lack of support for iWARP in tests 00:00:42.651 [Pipeline] withCredentials 00:00:42.663 > git --version # timeout=10 00:00:42.677 > git --version # 'git version 2.39.2' 00:00:42.697 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:42.699 [Pipeline] { 00:00:42.708 [Pipeline] retry 00:00:42.710 [Pipeline] { 00:00:42.726 [Pipeline] sh 00:00:43.013 + git ls-remote http://dpdk.org/git/dpdk-stable v23.11 00:00:47.232 [Pipeline] } 00:00:47.251 [Pipeline] // retry 00:00:47.256 [Pipeline] } 00:00:47.273 [Pipeline] // withCredentials 00:00:47.283 [Pipeline] httpRequest 00:00:47.722 [Pipeline] echo 00:00:47.724 Sorcerer 10.211.164.101 is alive 00:00:47.733 [Pipeline] retry 00:00:47.735 [Pipeline] { 00:00:47.749 [Pipeline] httpRequest 00:00:47.754 HttpMethod: GET 00:00:47.755 URL: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:00:47.755 Sending request to url: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:00:47.761 Response Code: HTTP/1.1 200 OK 00:00:47.762 Success: Status code 200 is in the accepted range: 200,404 00:00:47.762 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:01:35.906 [Pipeline] } 00:01:35.920 [Pipeline] // retry 00:01:35.931 [Pipeline] sh 00:01:36.208 + tar --no-same-owner -xf dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:01:37.600 [Pipeline] sh 00:01:37.884 + git -C dpdk log --oneline -n5 00:01:37.884 eeb0605f11 version: 23.11.0 00:01:37.884 238778122a doc: update release notes for 23.11 00:01:37.884 46aa6b3cfc doc: fix description of RSS features 00:01:37.884 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:01:37.884 7e421ae345 devtools: support skipping forbid rule check 00:01:37.904 [Pipeline] writeFile 00:01:37.924 [Pipeline] sh 00:01:38.214 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:38.225 [Pipeline] sh 00:01:38.508 + cat autorun-spdk.conf 00:01:38.509 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:38.509 SPDK_RUN_ASAN=1 00:01:38.509 SPDK_RUN_UBSAN=1 00:01:38.509 SPDK_TEST_RAID=1 00:01:38.509 SPDK_TEST_NATIVE_DPDK=v23.11 00:01:38.509 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:38.509 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:38.524 RUN_NIGHTLY=1 00:01:38.526 [Pipeline] } 00:01:38.540 [Pipeline] // stage 00:01:38.554 [Pipeline] stage 00:01:38.556 [Pipeline] { (Run VM) 00:01:38.568 [Pipeline] sh 00:01:38.889 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:38.889 + echo 'Start stage prepare_nvme.sh' 00:01:38.889 Start stage prepare_nvme.sh 00:01:38.889 + [[ -n 3 ]] 00:01:38.889 + disk_prefix=ex3 00:01:38.889 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:38.889 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:38.889 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:38.889 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:38.889 ++ SPDK_RUN_ASAN=1 00:01:38.889 ++ SPDK_RUN_UBSAN=1 00:01:38.889 ++ SPDK_TEST_RAID=1 00:01:38.889 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:01:38.889 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:38.889 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:38.889 ++ RUN_NIGHTLY=1 00:01:38.889 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:38.889 + nvme_files=() 00:01:38.889 + declare -A nvme_files 00:01:38.889 + backend_dir=/var/lib/libvirt/images/backends 00:01:38.889 + nvme_files['nvme.img']=5G 00:01:38.889 + nvme_files['nvme-cmb.img']=5G 00:01:38.889 + nvme_files['nvme-multi0.img']=4G 00:01:38.889 + nvme_files['nvme-multi1.img']=4G 00:01:38.889 + nvme_files['nvme-multi2.img']=4G 00:01:38.889 + nvme_files['nvme-openstack.img']=8G 00:01:38.889 + nvme_files['nvme-zns.img']=5G 00:01:38.889 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:38.889 + (( SPDK_TEST_FTL == 1 )) 00:01:38.889 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:38.889 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi2.img -s 4G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-cmb.img -s 5G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-openstack.img -s 8G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-zns.img -s 5G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi1.img -s 4G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi0.img -s 4G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:38.889 + for nvme in "${!nvme_files[@]}" 00:01:38.889 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme.img -s 5G 00:01:38.889 Formatting '/var/lib/libvirt/images/backends/ex3-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:38.889 ++ sudo grep -rl ex3-nvme.img /etc/libvirt/qemu 00:01:39.149 + echo 'End stage prepare_nvme.sh' 00:01:39.149 End stage prepare_nvme.sh 00:01:39.161 [Pipeline] sh 00:01:39.444 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:39.444 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex3-nvme.img -b /var/lib/libvirt/images/backends/ex3-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex3-nvme-multi1.img:/var/lib/libvirt/images/backends/ex3-nvme-multi2.img -H -a -v -f fedora39 00:01:39.444 00:01:39.444 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:39.444 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:39.444 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:39.444 HELP=0 00:01:39.444 DRY_RUN=0 00:01:39.444 NVME_FILE=/var/lib/libvirt/images/backends/ex3-nvme.img,/var/lib/libvirt/images/backends/ex3-nvme-multi0.img, 00:01:39.444 NVME_DISKS_TYPE=nvme,nvme, 00:01:39.444 NVME_AUTO_CREATE=0 00:01:39.444 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex3-nvme-multi1.img:/var/lib/libvirt/images/backends/ex3-nvme-multi2.img, 00:01:39.444 NVME_CMB=,, 00:01:39.444 NVME_PMR=,, 00:01:39.444 NVME_ZNS=,, 00:01:39.444 NVME_MS=,, 00:01:39.444 NVME_FDP=,, 00:01:39.444 SPDK_VAGRANT_DISTRO=fedora39 00:01:39.444 SPDK_VAGRANT_VMCPU=10 00:01:39.444 SPDK_VAGRANT_VMRAM=12288 00:01:39.444 SPDK_VAGRANT_PROVIDER=libvirt 00:01:39.444 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:39.445 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:39.445 SPDK_OPENSTACK_NETWORK=0 00:01:39.445 VAGRANT_PACKAGE_BOX=0 00:01:39.445 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:39.445 FORCE_DISTRO=true 00:01:39.445 VAGRANT_BOX_VERSION= 00:01:39.445 EXTRA_VAGRANTFILES= 00:01:39.445 NIC_MODEL=virtio 00:01:39.445 00:01:39.445 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:39.445 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:41.354 Bringing machine 'default' up with 'libvirt' provider... 00:01:41.923 ==> default: Creating image (snapshot of base box volume). 00:01:42.184 ==> default: Creating domain with the following settings... 00:01:42.184 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1730199047_85392fc672ac4fe8897e 00:01:42.184 ==> default: -- Domain type: kvm 00:01:42.184 ==> default: -- Cpus: 10 00:01:42.184 ==> default: -- Feature: acpi 00:01:42.184 ==> default: -- Feature: apic 00:01:42.184 ==> default: -- Feature: pae 00:01:42.184 ==> default: -- Memory: 12288M 00:01:42.184 ==> default: -- Memory Backing: hugepages: 00:01:42.184 ==> default: -- Management MAC: 00:01:42.184 ==> default: -- Loader: 00:01:42.184 ==> default: -- Nvram: 00:01:42.184 ==> default: -- Base box: spdk/fedora39 00:01:42.184 ==> default: -- Storage pool: default 00:01:42.184 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1730199047_85392fc672ac4fe8897e.img (20G) 00:01:42.184 ==> default: -- Volume Cache: default 00:01:42.184 ==> default: -- Kernel: 00:01:42.184 ==> default: -- Initrd: 00:01:42.184 ==> default: -- Graphics Type: vnc 00:01:42.184 ==> default: -- Graphics Port: -1 00:01:42.184 ==> default: -- Graphics IP: 127.0.0.1 00:01:42.184 ==> default: -- Graphics Password: Not defined 00:01:42.184 ==> default: -- Video Type: cirrus 00:01:42.184 ==> default: -- Video VRAM: 9216 00:01:42.184 ==> default: -- Sound Type: 00:01:42.184 ==> default: -- Keymap: en-us 00:01:42.184 ==> default: -- TPM Path: 00:01:42.184 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:42.184 ==> default: -- Command line args: 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:42.184 ==> default: -> value=-drive, 00:01:42.184 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme.img,if=none,id=nvme-0-drive0, 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:42.184 ==> default: -> value=-drive, 00:01:42.184 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:42.184 ==> default: -> value=-drive, 00:01:42.184 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:42.184 ==> default: -> value=-drive, 00:01:42.184 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:42.184 ==> default: -> value=-device, 00:01:42.184 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:42.184 ==> default: Creating shared folders metadata... 00:01:42.184 ==> default: Starting domain. 00:01:43.564 ==> default: Waiting for domain to get an IP address... 00:02:01.668 ==> default: Waiting for SSH to become available... 00:02:01.668 ==> default: Configuring and enabling network interfaces... 00:02:06.961 default: SSH address: 192.168.121.183:22 00:02:06.961 default: SSH username: vagrant 00:02:06.961 default: SSH auth method: private key 00:02:09.505 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:02:17.636 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:24.214 ==> default: Mounting SSHFS shared folder... 00:02:25.595 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:25.595 ==> default: Checking Mount.. 00:02:27.544 ==> default: Folder Successfully Mounted! 00:02:27.544 ==> default: Running provisioner: file... 00:02:28.500 default: ~/.gitconfig => .gitconfig 00:02:29.070 00:02:29.070 SUCCESS! 00:02:29.070 00:02:29.070 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:29.070 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:29.070 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:29.070 00:02:29.081 [Pipeline] } 00:02:29.101 [Pipeline] // stage 00:02:29.111 [Pipeline] dir 00:02:29.112 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:29.114 [Pipeline] { 00:02:29.129 [Pipeline] catchError 00:02:29.131 [Pipeline] { 00:02:29.146 [Pipeline] sh 00:02:29.436 + vagrant ssh-config --host vagrant 00:02:29.436 + + sed -ne /^Host/,$p 00:02:29.436 tee ssh_conf 00:02:32.006 Host vagrant 00:02:32.006 HostName 192.168.121.183 00:02:32.006 User vagrant 00:02:32.006 Port 22 00:02:32.006 UserKnownHostsFile /dev/null 00:02:32.006 StrictHostKeyChecking no 00:02:32.006 PasswordAuthentication no 00:02:32.006 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:32.006 IdentitiesOnly yes 00:02:32.006 LogLevel FATAL 00:02:32.006 ForwardAgent yes 00:02:32.006 ForwardX11 yes 00:02:32.006 00:02:32.021 [Pipeline] withEnv 00:02:32.024 [Pipeline] { 00:02:32.039 [Pipeline] sh 00:02:32.324 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:32.324 source /etc/os-release 00:02:32.324 [[ -e /image.version ]] && img=$(< /image.version) 00:02:32.324 # Minimal, systemd-like check. 00:02:32.324 if [[ -e /.dockerenv ]]; then 00:02:32.324 # Clear garbage from the node's name: 00:02:32.324 # agt-er_autotest_547-896 -> autotest_547-896 00:02:32.324 # $HOSTNAME is the actual container id 00:02:32.324 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:32.324 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:32.324 # We can assume this is a mount from a host where container is running, 00:02:32.324 # so fetch its hostname to easily identify the target swarm worker. 00:02:32.324 container="$(< /etc/hostname) ($agent)" 00:02:32.324 else 00:02:32.324 # Fallback 00:02:32.324 container=$agent 00:02:32.324 fi 00:02:32.324 fi 00:02:32.324 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:32.324 00:02:32.596 [Pipeline] } 00:02:32.610 [Pipeline] // withEnv 00:02:32.618 [Pipeline] setCustomBuildProperty 00:02:32.631 [Pipeline] stage 00:02:32.633 [Pipeline] { (Tests) 00:02:32.646 [Pipeline] sh 00:02:32.925 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:33.201 [Pipeline] sh 00:02:33.482 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:33.754 [Pipeline] timeout 00:02:33.755 Timeout set to expire in 1 hr 30 min 00:02:33.756 [Pipeline] { 00:02:33.770 [Pipeline] sh 00:02:34.048 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:34.614 HEAD is now at 12fc2abf1 test: Remove autopackage.sh 00:02:34.625 [Pipeline] sh 00:02:34.951 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:35.225 [Pipeline] sh 00:02:35.507 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:35.784 [Pipeline] sh 00:02:36.071 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:36.331 ++ readlink -f spdk_repo 00:02:36.331 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:36.331 + [[ -n /home/vagrant/spdk_repo ]] 00:02:36.331 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:36.331 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:36.331 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:36.331 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:36.331 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:36.331 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:36.331 + cd /home/vagrant/spdk_repo 00:02:36.331 + source /etc/os-release 00:02:36.331 ++ NAME='Fedora Linux' 00:02:36.331 ++ VERSION='39 (Cloud Edition)' 00:02:36.331 ++ ID=fedora 00:02:36.331 ++ VERSION_ID=39 00:02:36.331 ++ VERSION_CODENAME= 00:02:36.331 ++ PLATFORM_ID=platform:f39 00:02:36.331 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:36.331 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:36.331 ++ LOGO=fedora-logo-icon 00:02:36.331 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:36.331 ++ HOME_URL=https://fedoraproject.org/ 00:02:36.331 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:36.331 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:36.331 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:36.331 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:36.331 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:36.331 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:36.331 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:36.331 ++ SUPPORT_END=2024-11-12 00:02:36.331 ++ VARIANT='Cloud Edition' 00:02:36.331 ++ VARIANT_ID=cloud 00:02:36.331 + uname -a 00:02:36.331 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:36.331 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:36.899 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:36.899 Hugepages 00:02:36.899 node hugesize free / total 00:02:36.899 node0 1048576kB 0 / 0 00:02:36.899 node0 2048kB 0 / 0 00:02:36.899 00:02:36.899 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:36.899 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:36.899 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:02:36.899 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:02:36.899 + rm -f /tmp/spdk-ld-path 00:02:36.899 + source autorun-spdk.conf 00:02:36.899 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:36.899 ++ SPDK_RUN_ASAN=1 00:02:36.899 ++ SPDK_RUN_UBSAN=1 00:02:36.899 ++ SPDK_TEST_RAID=1 00:02:36.899 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:02:36.900 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:36.900 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:36.900 ++ RUN_NIGHTLY=1 00:02:36.900 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:36.900 + [[ -n '' ]] 00:02:36.900 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:37.159 + for M in /var/spdk/build-*-manifest.txt 00:02:37.159 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:37.159 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:37.159 + for M in /var/spdk/build-*-manifest.txt 00:02:37.159 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:37.159 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:37.159 + for M in /var/spdk/build-*-manifest.txt 00:02:37.159 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:37.159 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:37.159 ++ uname 00:02:37.159 + [[ Linux == \L\i\n\u\x ]] 00:02:37.159 + sudo dmesg -T 00:02:37.159 + sudo dmesg --clear 00:02:37.159 + dmesg_pid=6152 00:02:37.159 + sudo dmesg -Tw 00:02:37.159 + [[ Fedora Linux == FreeBSD ]] 00:02:37.159 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:37.159 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:37.159 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:37.159 + [[ -x /usr/src/fio-static/fio ]] 00:02:37.159 + export FIO_BIN=/usr/src/fio-static/fio 00:02:37.159 + FIO_BIN=/usr/src/fio-static/fio 00:02:37.159 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:37.159 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:37.159 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:37.159 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:37.159 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:37.159 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:37.159 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:37.159 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:37.159 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:37.419 10:51:42 -- common/autotest_common.sh@1690 -- $ [[ n == y ]] 00:02:37.419 10:51:42 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v23.11 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:37.419 10:51:42 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:02:37.419 10:51:42 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:02:37.419 10:51:42 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:37.419 10:51:42 -- common/autotest_common.sh@1690 -- $ [[ n == y ]] 00:02:37.419 10:51:42 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:37.419 10:51:42 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:37.419 10:51:42 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:37.419 10:51:42 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:37.419 10:51:42 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:37.419 10:51:42 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:37.420 10:51:42 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:37.420 10:51:42 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:37.420 10:51:42 -- paths/export.sh@5 -- $ export PATH 00:02:37.420 10:51:42 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:37.420 10:51:42 -- common/autobuild_common.sh@485 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:37.420 10:51:42 -- common/autobuild_common.sh@486 -- $ date +%s 00:02:37.420 10:51:42 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1730199102.XXXXXX 00:02:37.420 10:51:42 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1730199102.ScTQLM 00:02:37.420 10:51:42 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:02:37.420 10:51:42 -- common/autobuild_common.sh@492 -- $ '[' -n v23.11 ']' 00:02:37.420 10:51:42 -- common/autobuild_common.sh@493 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:37.420 10:51:42 -- common/autobuild_common.sh@493 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:37.420 10:51:42 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:37.420 10:51:42 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:37.420 10:51:42 -- common/autobuild_common.sh@502 -- $ get_config_params 00:02:37.420 10:51:42 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:02:37.420 10:51:42 -- common/autotest_common.sh@10 -- $ set +x 00:02:37.420 10:51:42 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:37.420 10:51:42 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:02:37.420 10:51:42 -- pm/common@17 -- $ local monitor 00:02:37.420 10:51:42 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:37.420 10:51:42 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:37.420 10:51:42 -- pm/common@25 -- $ sleep 1 00:02:37.420 10:51:42 -- pm/common@21 -- $ date +%s 00:02:37.420 10:51:42 -- pm/common@21 -- $ date +%s 00:02:37.420 10:51:42 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1730199102 00:02:37.420 10:51:42 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1730199102 00:02:37.420 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1730199102_collect-vmstat.pm.log 00:02:37.420 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1730199102_collect-cpu-load.pm.log 00:02:38.357 10:51:43 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:02:38.357 10:51:43 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:38.357 10:51:43 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:38.357 10:51:43 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:38.357 10:51:43 -- spdk/autobuild.sh@16 -- $ date -u 00:02:38.357 Tue Oct 29 10:51:43 AM UTC 2024 00:02:38.357 10:51:43 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:38.615 v25.01-pre-123-g12fc2abf1 00:02:38.615 10:51:43 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:38.615 10:51:43 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:38.615 10:51:43 -- common/autotest_common.sh@1103 -- $ '[' 3 -le 1 ']' 00:02:38.615 10:51:43 -- common/autotest_common.sh@1109 -- $ xtrace_disable 00:02:38.615 10:51:43 -- common/autotest_common.sh@10 -- $ set +x 00:02:38.616 ************************************ 00:02:38.616 START TEST asan 00:02:38.616 ************************************ 00:02:38.616 using asan 00:02:38.616 10:51:43 asan -- common/autotest_common.sh@1127 -- $ echo 'using asan' 00:02:38.616 00:02:38.616 real 0m0.000s 00:02:38.616 user 0m0.000s 00:02:38.616 sys 0m0.000s 00:02:38.616 10:51:43 asan -- common/autotest_common.sh@1128 -- $ xtrace_disable 00:02:38.616 10:51:43 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:38.616 ************************************ 00:02:38.616 END TEST asan 00:02:38.616 ************************************ 00:02:38.616 10:51:43 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:38.616 10:51:43 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:38.616 10:51:43 -- common/autotest_common.sh@1103 -- $ '[' 3 -le 1 ']' 00:02:38.616 10:51:43 -- common/autotest_common.sh@1109 -- $ xtrace_disable 00:02:38.616 10:51:43 -- common/autotest_common.sh@10 -- $ set +x 00:02:38.616 ************************************ 00:02:38.616 START TEST ubsan 00:02:38.616 ************************************ 00:02:38.616 using ubsan 00:02:38.616 10:51:43 ubsan -- common/autotest_common.sh@1127 -- $ echo 'using ubsan' 00:02:38.616 00:02:38.616 real 0m0.000s 00:02:38.616 user 0m0.000s 00:02:38.616 sys 0m0.000s 00:02:38.616 10:51:43 ubsan -- common/autotest_common.sh@1128 -- $ xtrace_disable 00:02:38.616 10:51:43 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:38.616 ************************************ 00:02:38.616 END TEST ubsan 00:02:38.616 ************************************ 00:02:38.616 10:51:43 -- spdk/autobuild.sh@27 -- $ '[' -n v23.11 ']' 00:02:38.616 10:51:43 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:38.616 10:51:43 -- common/autobuild_common.sh@442 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:38.616 10:51:43 -- common/autotest_common.sh@1103 -- $ '[' 2 -le 1 ']' 00:02:38.616 10:51:43 -- common/autotest_common.sh@1109 -- $ xtrace_disable 00:02:38.616 10:51:43 -- common/autotest_common.sh@10 -- $ set +x 00:02:38.616 ************************************ 00:02:38.616 START TEST build_native_dpdk 00:02:38.616 ************************************ 00:02:38.616 10:51:44 build_native_dpdk -- common/autotest_common.sh@1127 -- $ _build_native_dpdk 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:38.616 eeb0605f11 version: 23.11.0 00:02:38.616 238778122a doc: update release notes for 23.11 00:02:38.616 46aa6b3cfc doc: fix description of RSS features 00:02:38.616 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:02:38.616 7e421ae345 devtools: support skipping forbid rule check 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=23.11.0 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@100 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base") 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@102 -- $ local mlx5_libs_added=n 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@139 -- $ [[ 0 -eq 1 ]] 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@167 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@168 -- $ uname -s 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@168 -- $ '[' Linux = Linux ']' 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@169 -- $ lt 23.11.0 21.11.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 21.11.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@173 -- $ patch -p1 00:02:38.616 patching file config/rte_config.h 00:02:38.616 Hunk #1 succeeded at 60 (offset 1 line). 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 23.11.0 24.07.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 24.07.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@177 -- $ patch -p1 00:02:38.616 patching file lib/pcapng/rte_pcapng.c 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@179 -- $ ge 23.11.0 24.07.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 23.11.0 '>=' 24.07.0 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:38.616 10:51:44 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@183 -- $ dpdk_kmods=false 00:02:38.616 10:51:44 build_native_dpdk -- common/autobuild_common.sh@184 -- $ uname -s 00:02:38.875 10:51:44 build_native_dpdk -- common/autobuild_common.sh@184 -- $ '[' Linux = FreeBSD ']' 00:02:38.875 10:51:44 build_native_dpdk -- common/autobuild_common.sh@188 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base 00:02:38.875 10:51:44 build_native_dpdk -- common/autobuild_common.sh@188 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:45.444 The Meson build system 00:02:45.444 Version: 1.5.0 00:02:45.444 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:45.444 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:45.444 Build type: native build 00:02:45.444 Program cat found: YES (/usr/bin/cat) 00:02:45.444 Project name: DPDK 00:02:45.444 Project version: 23.11.0 00:02:45.444 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:45.444 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:45.444 Host machine cpu family: x86_64 00:02:45.444 Host machine cpu: x86_64 00:02:45.444 Message: ## Building in Developer Mode ## 00:02:45.444 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:45.444 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:45.444 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:45.444 Program python3 found: YES (/usr/bin/python3) 00:02:45.444 Program cat found: YES (/usr/bin/cat) 00:02:45.444 config/meson.build:113: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:45.444 Compiler for C supports arguments -march=native: YES 00:02:45.444 Checking for size of "void *" : 8 00:02:45.444 Checking for size of "void *" : 8 (cached) 00:02:45.444 Library m found: YES 00:02:45.444 Library numa found: YES 00:02:45.444 Has header "numaif.h" : YES 00:02:45.444 Library fdt found: NO 00:02:45.444 Library execinfo found: NO 00:02:45.444 Has header "execinfo.h" : YES 00:02:45.444 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:45.444 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:45.444 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:45.444 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:45.444 Run-time dependency openssl found: YES 3.1.1 00:02:45.444 Run-time dependency libpcap found: YES 1.10.4 00:02:45.444 Has header "pcap.h" with dependency libpcap: YES 00:02:45.444 Compiler for C supports arguments -Wcast-qual: YES 00:02:45.444 Compiler for C supports arguments -Wdeprecated: YES 00:02:45.444 Compiler for C supports arguments -Wformat: YES 00:02:45.444 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:45.444 Compiler for C supports arguments -Wformat-security: NO 00:02:45.444 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:45.444 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:45.444 Compiler for C supports arguments -Wnested-externs: YES 00:02:45.444 Compiler for C supports arguments -Wold-style-definition: YES 00:02:45.444 Compiler for C supports arguments -Wpointer-arith: YES 00:02:45.444 Compiler for C supports arguments -Wsign-compare: YES 00:02:45.444 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:45.444 Compiler for C supports arguments -Wundef: YES 00:02:45.444 Compiler for C supports arguments -Wwrite-strings: YES 00:02:45.444 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:45.444 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:45.444 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:45.444 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:45.444 Program objdump found: YES (/usr/bin/objdump) 00:02:45.444 Compiler for C supports arguments -mavx512f: YES 00:02:45.444 Checking if "AVX512 checking" compiles: YES 00:02:45.444 Fetching value of define "__SSE4_2__" : 1 00:02:45.444 Fetching value of define "__AES__" : 1 00:02:45.444 Fetching value of define "__AVX__" : 1 00:02:45.444 Fetching value of define "__AVX2__" : 1 00:02:45.444 Fetching value of define "__AVX512BW__" : 1 00:02:45.444 Fetching value of define "__AVX512CD__" : 1 00:02:45.444 Fetching value of define "__AVX512DQ__" : 1 00:02:45.444 Fetching value of define "__AVX512F__" : 1 00:02:45.444 Fetching value of define "__AVX512VL__" : 1 00:02:45.444 Fetching value of define "__PCLMUL__" : 1 00:02:45.444 Fetching value of define "__RDRND__" : 1 00:02:45.444 Fetching value of define "__RDSEED__" : 1 00:02:45.444 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:45.444 Fetching value of define "__znver1__" : (undefined) 00:02:45.444 Fetching value of define "__znver2__" : (undefined) 00:02:45.444 Fetching value of define "__znver3__" : (undefined) 00:02:45.444 Fetching value of define "__znver4__" : (undefined) 00:02:45.444 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:45.444 Message: lib/log: Defining dependency "log" 00:02:45.444 Message: lib/kvargs: Defining dependency "kvargs" 00:02:45.444 Message: lib/telemetry: Defining dependency "telemetry" 00:02:45.444 Checking for function "getentropy" : NO 00:02:45.444 Message: lib/eal: Defining dependency "eal" 00:02:45.444 Message: lib/ring: Defining dependency "ring" 00:02:45.444 Message: lib/rcu: Defining dependency "rcu" 00:02:45.444 Message: lib/mempool: Defining dependency "mempool" 00:02:45.444 Message: lib/mbuf: Defining dependency "mbuf" 00:02:45.444 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:45.444 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:45.444 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:45.444 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:45.444 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:45.444 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:45.444 Compiler for C supports arguments -mpclmul: YES 00:02:45.444 Compiler for C supports arguments -maes: YES 00:02:45.444 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:45.444 Compiler for C supports arguments -mavx512bw: YES 00:02:45.444 Compiler for C supports arguments -mavx512dq: YES 00:02:45.444 Compiler for C supports arguments -mavx512vl: YES 00:02:45.445 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:45.445 Compiler for C supports arguments -mavx2: YES 00:02:45.445 Compiler for C supports arguments -mavx: YES 00:02:45.445 Message: lib/net: Defining dependency "net" 00:02:45.445 Message: lib/meter: Defining dependency "meter" 00:02:45.445 Message: lib/ethdev: Defining dependency "ethdev" 00:02:45.445 Message: lib/pci: Defining dependency "pci" 00:02:45.445 Message: lib/cmdline: Defining dependency "cmdline" 00:02:45.445 Message: lib/metrics: Defining dependency "metrics" 00:02:45.445 Message: lib/hash: Defining dependency "hash" 00:02:45.445 Message: lib/timer: Defining dependency "timer" 00:02:45.445 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:45.445 Message: lib/acl: Defining dependency "acl" 00:02:45.445 Message: lib/bbdev: Defining dependency "bbdev" 00:02:45.445 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:45.445 Run-time dependency libelf found: YES 0.191 00:02:45.445 Message: lib/bpf: Defining dependency "bpf" 00:02:45.445 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:45.445 Message: lib/compressdev: Defining dependency "compressdev" 00:02:45.445 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:45.445 Message: lib/distributor: Defining dependency "distributor" 00:02:45.445 Message: lib/dmadev: Defining dependency "dmadev" 00:02:45.445 Message: lib/efd: Defining dependency "efd" 00:02:45.445 Message: lib/eventdev: Defining dependency "eventdev" 00:02:45.445 Message: lib/dispatcher: Defining dependency "dispatcher" 00:02:45.445 Message: lib/gpudev: Defining dependency "gpudev" 00:02:45.445 Message: lib/gro: Defining dependency "gro" 00:02:45.445 Message: lib/gso: Defining dependency "gso" 00:02:45.445 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:45.445 Message: lib/jobstats: Defining dependency "jobstats" 00:02:45.445 Message: lib/latencystats: Defining dependency "latencystats" 00:02:45.445 Message: lib/lpm: Defining dependency "lpm" 00:02:45.445 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:45.445 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:45.445 Message: lib/member: Defining dependency "member" 00:02:45.445 Message: lib/pcapng: Defining dependency "pcapng" 00:02:45.445 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:45.445 Message: lib/power: Defining dependency "power" 00:02:45.445 Message: lib/rawdev: Defining dependency "rawdev" 00:02:45.445 Message: lib/regexdev: Defining dependency "regexdev" 00:02:45.445 Message: lib/mldev: Defining dependency "mldev" 00:02:45.445 Message: lib/rib: Defining dependency "rib" 00:02:45.445 Message: lib/reorder: Defining dependency "reorder" 00:02:45.445 Message: lib/sched: Defining dependency "sched" 00:02:45.445 Message: lib/security: Defining dependency "security" 00:02:45.445 Message: lib/stack: Defining dependency "stack" 00:02:45.445 Has header "linux/userfaultfd.h" : YES 00:02:45.445 Has header "linux/vduse.h" : YES 00:02:45.445 Message: lib/vhost: Defining dependency "vhost" 00:02:45.445 Message: lib/ipsec: Defining dependency "ipsec" 00:02:45.445 Message: lib/pdcp: Defining dependency "pdcp" 00:02:45.445 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:45.445 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:45.445 Message: lib/fib: Defining dependency "fib" 00:02:45.445 Message: lib/port: Defining dependency "port" 00:02:45.445 Message: lib/pdump: Defining dependency "pdump" 00:02:45.445 Message: lib/table: Defining dependency "table" 00:02:45.445 Message: lib/pipeline: Defining dependency "pipeline" 00:02:45.445 Message: lib/graph: Defining dependency "graph" 00:02:45.445 Message: lib/node: Defining dependency "node" 00:02:45.445 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:45.445 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:45.445 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:46.823 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:46.823 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:46.823 Compiler for C supports arguments -Wno-unused-value: YES 00:02:46.823 Compiler for C supports arguments -Wno-format: YES 00:02:46.823 Compiler for C supports arguments -Wno-format-security: YES 00:02:46.823 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:46.823 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:46.823 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:46.823 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:46.823 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:46.823 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:46.823 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:46.823 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:46.823 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:46.823 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:46.823 Has header "sys/epoll.h" : YES 00:02:46.823 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:46.823 Configuring doxy-api-html.conf using configuration 00:02:46.823 Configuring doxy-api-man.conf using configuration 00:02:46.823 Program mandb found: YES (/usr/bin/mandb) 00:02:46.823 Program sphinx-build found: NO 00:02:46.823 Configuring rte_build_config.h using configuration 00:02:46.823 Message: 00:02:46.823 ================= 00:02:46.823 Applications Enabled 00:02:46.823 ================= 00:02:46.823 00:02:46.823 apps: 00:02:46.823 dumpcap, graph, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, 00:02:46.823 test-crypto-perf, test-dma-perf, test-eventdev, test-fib, test-flow-perf, test-gpudev, test-mldev, test-pipeline, 00:02:46.823 test-pmd, test-regex, test-sad, test-security-perf, 00:02:46.823 00:02:46.823 Message: 00:02:46.823 ================= 00:02:46.823 Libraries Enabled 00:02:46.823 ================= 00:02:46.823 00:02:46.823 libs: 00:02:46.823 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:02:46.823 net, meter, ethdev, pci, cmdline, metrics, hash, timer, 00:02:46.823 acl, bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, 00:02:46.823 dmadev, efd, eventdev, dispatcher, gpudev, gro, gso, ip_frag, 00:02:46.823 jobstats, latencystats, lpm, member, pcapng, power, rawdev, regexdev, 00:02:46.823 mldev, rib, reorder, sched, security, stack, vhost, ipsec, 00:02:46.823 pdcp, fib, port, pdump, table, pipeline, graph, node, 00:02:46.823 00:02:46.823 00:02:46.823 Message: 00:02:46.823 =============== 00:02:46.823 Drivers Enabled 00:02:46.823 =============== 00:02:46.823 00:02:46.823 common: 00:02:46.823 00:02:46.823 bus: 00:02:46.823 pci, vdev, 00:02:46.823 mempool: 00:02:46.823 ring, 00:02:46.823 dma: 00:02:46.823 00:02:46.823 net: 00:02:46.823 i40e, 00:02:46.823 raw: 00:02:46.823 00:02:46.823 crypto: 00:02:46.823 00:02:46.823 compress: 00:02:46.823 00:02:46.823 regex: 00:02:46.823 00:02:46.823 ml: 00:02:46.823 00:02:46.823 vdpa: 00:02:46.823 00:02:46.823 event: 00:02:46.823 00:02:46.823 baseband: 00:02:46.823 00:02:46.823 gpu: 00:02:46.823 00:02:46.823 00:02:46.823 Message: 00:02:46.823 ================= 00:02:46.823 Content Skipped 00:02:46.823 ================= 00:02:46.823 00:02:46.823 apps: 00:02:46.823 00:02:46.823 libs: 00:02:46.823 00:02:46.823 drivers: 00:02:46.823 common/cpt: not in enabled drivers build config 00:02:46.823 common/dpaax: not in enabled drivers build config 00:02:46.823 common/iavf: not in enabled drivers build config 00:02:46.823 common/idpf: not in enabled drivers build config 00:02:46.823 common/mvep: not in enabled drivers build config 00:02:46.823 common/octeontx: not in enabled drivers build config 00:02:46.823 bus/auxiliary: not in enabled drivers build config 00:02:46.823 bus/cdx: not in enabled drivers build config 00:02:46.823 bus/dpaa: not in enabled drivers build config 00:02:46.823 bus/fslmc: not in enabled drivers build config 00:02:46.823 bus/ifpga: not in enabled drivers build config 00:02:46.823 bus/platform: not in enabled drivers build config 00:02:46.823 bus/vmbus: not in enabled drivers build config 00:02:46.823 common/cnxk: not in enabled drivers build config 00:02:46.823 common/mlx5: not in enabled drivers build config 00:02:46.823 common/nfp: not in enabled drivers build config 00:02:46.823 common/qat: not in enabled drivers build config 00:02:46.823 common/sfc_efx: not in enabled drivers build config 00:02:46.823 mempool/bucket: not in enabled drivers build config 00:02:46.823 mempool/cnxk: not in enabled drivers build config 00:02:46.823 mempool/dpaa: not in enabled drivers build config 00:02:46.823 mempool/dpaa2: not in enabled drivers build config 00:02:46.823 mempool/octeontx: not in enabled drivers build config 00:02:46.823 mempool/stack: not in enabled drivers build config 00:02:46.823 dma/cnxk: not in enabled drivers build config 00:02:46.823 dma/dpaa: not in enabled drivers build config 00:02:46.823 dma/dpaa2: not in enabled drivers build config 00:02:46.823 dma/hisilicon: not in enabled drivers build config 00:02:46.823 dma/idxd: not in enabled drivers build config 00:02:46.823 dma/ioat: not in enabled drivers build config 00:02:46.823 dma/skeleton: not in enabled drivers build config 00:02:46.823 net/af_packet: not in enabled drivers build config 00:02:46.823 net/af_xdp: not in enabled drivers build config 00:02:46.823 net/ark: not in enabled drivers build config 00:02:46.823 net/atlantic: not in enabled drivers build config 00:02:46.823 net/avp: not in enabled drivers build config 00:02:46.823 net/axgbe: not in enabled drivers build config 00:02:46.823 net/bnx2x: not in enabled drivers build config 00:02:46.823 net/bnxt: not in enabled drivers build config 00:02:46.823 net/bonding: not in enabled drivers build config 00:02:46.823 net/cnxk: not in enabled drivers build config 00:02:46.823 net/cpfl: not in enabled drivers build config 00:02:46.823 net/cxgbe: not in enabled drivers build config 00:02:46.823 net/dpaa: not in enabled drivers build config 00:02:46.823 net/dpaa2: not in enabled drivers build config 00:02:46.823 net/e1000: not in enabled drivers build config 00:02:46.823 net/ena: not in enabled drivers build config 00:02:46.823 net/enetc: not in enabled drivers build config 00:02:46.823 net/enetfec: not in enabled drivers build config 00:02:46.823 net/enic: not in enabled drivers build config 00:02:46.823 net/failsafe: not in enabled drivers build config 00:02:46.823 net/fm10k: not in enabled drivers build config 00:02:46.823 net/gve: not in enabled drivers build config 00:02:46.823 net/hinic: not in enabled drivers build config 00:02:46.823 net/hns3: not in enabled drivers build config 00:02:46.823 net/iavf: not in enabled drivers build config 00:02:46.823 net/ice: not in enabled drivers build config 00:02:46.823 net/idpf: not in enabled drivers build config 00:02:46.823 net/igc: not in enabled drivers build config 00:02:46.823 net/ionic: not in enabled drivers build config 00:02:46.823 net/ipn3ke: not in enabled drivers build config 00:02:46.823 net/ixgbe: not in enabled drivers build config 00:02:46.823 net/mana: not in enabled drivers build config 00:02:46.823 net/memif: not in enabled drivers build config 00:02:46.823 net/mlx4: not in enabled drivers build config 00:02:46.823 net/mlx5: not in enabled drivers build config 00:02:46.823 net/mvneta: not in enabled drivers build config 00:02:46.823 net/mvpp2: not in enabled drivers build config 00:02:46.823 net/netvsc: not in enabled drivers build config 00:02:46.823 net/nfb: not in enabled drivers build config 00:02:46.823 net/nfp: not in enabled drivers build config 00:02:46.823 net/ngbe: not in enabled drivers build config 00:02:46.823 net/null: not in enabled drivers build config 00:02:46.823 net/octeontx: not in enabled drivers build config 00:02:46.823 net/octeon_ep: not in enabled drivers build config 00:02:46.823 net/pcap: not in enabled drivers build config 00:02:46.824 net/pfe: not in enabled drivers build config 00:02:46.824 net/qede: not in enabled drivers build config 00:02:46.824 net/ring: not in enabled drivers build config 00:02:46.824 net/sfc: not in enabled drivers build config 00:02:46.824 net/softnic: not in enabled drivers build config 00:02:46.824 net/tap: not in enabled drivers build config 00:02:46.824 net/thunderx: not in enabled drivers build config 00:02:46.824 net/txgbe: not in enabled drivers build config 00:02:46.824 net/vdev_netvsc: not in enabled drivers build config 00:02:46.824 net/vhost: not in enabled drivers build config 00:02:46.824 net/virtio: not in enabled drivers build config 00:02:46.824 net/vmxnet3: not in enabled drivers build config 00:02:46.824 raw/cnxk_bphy: not in enabled drivers build config 00:02:46.824 raw/cnxk_gpio: not in enabled drivers build config 00:02:46.824 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:46.824 raw/ifpga: not in enabled drivers build config 00:02:46.824 raw/ntb: not in enabled drivers build config 00:02:46.824 raw/skeleton: not in enabled drivers build config 00:02:46.824 crypto/armv8: not in enabled drivers build config 00:02:46.824 crypto/bcmfs: not in enabled drivers build config 00:02:46.824 crypto/caam_jr: not in enabled drivers build config 00:02:46.824 crypto/ccp: not in enabled drivers build config 00:02:46.824 crypto/cnxk: not in enabled drivers build config 00:02:46.824 crypto/dpaa_sec: not in enabled drivers build config 00:02:46.824 crypto/dpaa2_sec: not in enabled drivers build config 00:02:46.824 crypto/ipsec_mb: not in enabled drivers build config 00:02:46.824 crypto/mlx5: not in enabled drivers build config 00:02:46.824 crypto/mvsam: not in enabled drivers build config 00:02:46.824 crypto/nitrox: not in enabled drivers build config 00:02:46.824 crypto/null: not in enabled drivers build config 00:02:46.824 crypto/octeontx: not in enabled drivers build config 00:02:46.824 crypto/openssl: not in enabled drivers build config 00:02:46.824 crypto/scheduler: not in enabled drivers build config 00:02:46.824 crypto/uadk: not in enabled drivers build config 00:02:46.824 crypto/virtio: not in enabled drivers build config 00:02:46.824 compress/isal: not in enabled drivers build config 00:02:46.824 compress/mlx5: not in enabled drivers build config 00:02:46.824 compress/octeontx: not in enabled drivers build config 00:02:46.824 compress/zlib: not in enabled drivers build config 00:02:46.824 regex/mlx5: not in enabled drivers build config 00:02:46.824 regex/cn9k: not in enabled drivers build config 00:02:46.824 ml/cnxk: not in enabled drivers build config 00:02:46.824 vdpa/ifc: not in enabled drivers build config 00:02:46.824 vdpa/mlx5: not in enabled drivers build config 00:02:46.824 vdpa/nfp: not in enabled drivers build config 00:02:46.824 vdpa/sfc: not in enabled drivers build config 00:02:46.824 event/cnxk: not in enabled drivers build config 00:02:46.824 event/dlb2: not in enabled drivers build config 00:02:46.824 event/dpaa: not in enabled drivers build config 00:02:46.824 event/dpaa2: not in enabled drivers build config 00:02:46.824 event/dsw: not in enabled drivers build config 00:02:46.824 event/opdl: not in enabled drivers build config 00:02:46.824 event/skeleton: not in enabled drivers build config 00:02:46.824 event/sw: not in enabled drivers build config 00:02:46.824 event/octeontx: not in enabled drivers build config 00:02:46.824 baseband/acc: not in enabled drivers build config 00:02:46.824 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:46.824 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:46.824 baseband/la12xx: not in enabled drivers build config 00:02:46.824 baseband/null: not in enabled drivers build config 00:02:46.824 baseband/turbo_sw: not in enabled drivers build config 00:02:46.824 gpu/cuda: not in enabled drivers build config 00:02:46.824 00:02:46.824 00:02:46.824 Build targets in project: 217 00:02:46.824 00:02:46.824 DPDK 23.11.0 00:02:46.824 00:02:46.824 User defined options 00:02:46.824 libdir : lib 00:02:46.824 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:46.824 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:46.824 c_link_args : 00:02:46.824 enable_docs : false 00:02:46.824 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:46.824 enable_kmods : false 00:02:46.824 machine : native 00:02:46.824 tests : false 00:02:46.824 00:02:46.824 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:46.824 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:47.083 10:51:52 build_native_dpdk -- common/autobuild_common.sh@192 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:47.084 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:47.084 [1/707] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:02:47.084 [2/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:47.084 [3/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:47.084 [4/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:47.084 [5/707] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:47.084 [6/707] Linking static target lib/librte_kvargs.a 00:02:47.343 [7/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:47.343 [8/707] Compiling C object lib/librte_log.a.p/log_log.c.o 00:02:47.343 [9/707] Linking static target lib/librte_log.a 00:02:47.343 [10/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:47.343 [11/707] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.602 [12/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:47.602 [13/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:47.602 [14/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:47.602 [15/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:47.602 [16/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:47.602 [17/707] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.861 [18/707] Linking target lib/librte_log.so.24.0 00:02:47.861 [19/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:47.861 [20/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:47.861 [21/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:47.861 [22/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:47.861 [23/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:48.120 [24/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:48.120 [25/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:48.120 [26/707] Generating symbol file lib/librte_log.so.24.0.p/librte_log.so.24.0.symbols 00:02:48.120 [27/707] Linking target lib/librte_kvargs.so.24.0 00:02:48.120 [28/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:48.120 [29/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:48.120 [30/707] Generating symbol file lib/librte_kvargs.so.24.0.p/librte_kvargs.so.24.0.symbols 00:02:48.120 [31/707] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:48.380 [32/707] Linking static target lib/librte_telemetry.a 00:02:48.380 [33/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:48.380 [34/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:48.380 [35/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:48.380 [36/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:48.380 [37/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:48.380 [38/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:48.380 [39/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:48.380 [40/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:48.639 [41/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:48.639 [42/707] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.639 [43/707] Linking target lib/librte_telemetry.so.24.0 00:02:48.639 [44/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:48.639 [45/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:48.899 [46/707] Generating symbol file lib/librte_telemetry.so.24.0.p/librte_telemetry.so.24.0.symbols 00:02:48.899 [47/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:48.899 [48/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:48.899 [49/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:48.899 [50/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:48.899 [51/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:48.899 [52/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:49.158 [53/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:49.158 [54/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:49.158 [55/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:49.158 [56/707] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:49.158 [57/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:49.158 [58/707] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:49.158 [59/707] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:49.418 [60/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:49.418 [61/707] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:49.418 [62/707] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:49.418 [63/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:49.418 [64/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:49.418 [65/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:49.418 [66/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:49.418 [67/707] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:49.418 [68/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:49.678 [69/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:49.678 [70/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:49.678 [71/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:49.678 [72/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:49.678 [73/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:49.937 [74/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:49.937 [75/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:49.937 [76/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:49.937 [77/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:49.937 [78/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:50.197 [79/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:50.197 [80/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:50.197 [81/707] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:50.197 [82/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:50.197 [83/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:50.197 [84/707] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:50.197 [85/707] Linking static target lib/librte_ring.a 00:02:50.456 [86/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:50.456 [87/707] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:50.456 [88/707] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.456 [89/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:50.456 [90/707] Linking static target lib/librte_eal.a 00:02:50.456 [91/707] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:50.714 [92/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:50.714 [93/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:50.714 [94/707] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:50.714 [95/707] Linking static target lib/librte_mempool.a 00:02:50.974 [96/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:50.974 [97/707] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:50.974 [98/707] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:50.974 [99/707] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:50.974 [100/707] Linking static target lib/librte_rcu.a 00:02:50.974 [101/707] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:50.974 [102/707] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:50.974 [103/707] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:50.974 [104/707] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:51.234 [105/707] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.234 [106/707] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.234 [107/707] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:51.234 [108/707] Linking static target lib/librte_net.a 00:02:51.494 [109/707] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:51.494 [110/707] Linking static target lib/librte_meter.a 00:02:51.494 [111/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:51.494 [112/707] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.494 [113/707] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:51.494 [114/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:51.495 [115/707] Linking static target lib/librte_mbuf.a 00:02:51.495 [116/707] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.755 [117/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:51.755 [118/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:52.015 [119/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:52.015 [120/707] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.015 [121/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:52.274 [122/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:02:52.533 [123/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:52.533 [124/707] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:52.533 [125/707] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:52.533 [126/707] Linking static target lib/librte_pci.a 00:02:52.533 [127/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:52.533 [128/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:52.533 [129/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:52.792 [130/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:52.793 [131/707] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.793 [132/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:52.793 [133/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:52.793 [134/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:52.793 [135/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:52.793 [136/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:52.793 [137/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:52.793 [138/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:52.793 [139/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:53.057 [140/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:53.057 [141/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:53.057 [142/707] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:53.057 [143/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:53.057 [144/707] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:53.057 [145/707] Linking static target lib/librte_cmdline.a 00:02:53.323 [146/707] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:53.323 [147/707] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:53.323 [148/707] Linking static target lib/librte_metrics.a 00:02:53.323 [149/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:53.323 [150/707] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:53.892 [151/707] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.892 [152/707] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:53.892 [153/707] Linking static target lib/librte_timer.a 00:02:53.892 [154/707] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:53.892 [155/707] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:54.151 [156/707] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.151 [157/707] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.151 [158/707] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:54.410 [159/707] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:54.410 [160/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:54.669 [161/707] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:54.927 [162/707] Linking static target lib/librte_bitratestats.a 00:02:54.927 [163/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:54.927 [164/707] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:54.927 [165/707] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:54.927 [166/707] Linking static target lib/librte_bbdev.a 00:02:54.927 [167/707] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:55.187 [168/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:55.447 [169/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:55.447 [170/707] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:55.706 [171/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:55.706 [172/707] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:55.706 [173/707] Linking static target lib/librte_ethdev.a 00:02:55.706 [174/707] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:55.706 [175/707] Linking static target lib/librte_hash.a 00:02:55.706 [176/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:55.706 [177/707] Compiling C object lib/acl/libavx2_tmp.a.p/acl_run_avx2.c.o 00:02:55.706 [178/707] Linking static target lib/acl/libavx2_tmp.a 00:02:55.964 [179/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:55.964 [180/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:56.224 [181/707] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:56.224 [182/707] Linking static target lib/librte_cfgfile.a 00:02:56.224 [183/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:56.224 [184/707] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.224 [185/707] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.483 [186/707] Linking target lib/librte_eal.so.24.0 00:02:56.483 [187/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:56.483 [188/707] Generating symbol file lib/librte_eal.so.24.0.p/librte_eal.so.24.0.symbols 00:02:56.483 [189/707] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.483 [190/707] Linking target lib/librte_ring.so.24.0 00:02:56.483 [191/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:56.483 [192/707] Linking target lib/librte_meter.so.24.0 00:02:56.483 [193/707] Linking target lib/librte_pci.so.24.0 00:02:56.743 [194/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:56.743 [195/707] Generating symbol file lib/librte_ring.so.24.0.p/librte_ring.so.24.0.symbols 00:02:56.743 [196/707] Linking target lib/librte_timer.so.24.0 00:02:56.743 [197/707] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:56.743 [198/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:56.743 [199/707] Generating symbol file lib/librte_pci.so.24.0.p/librte_pci.so.24.0.symbols 00:02:56.743 [200/707] Linking static target lib/librte_bpf.a 00:02:56.743 [201/707] Linking target lib/librte_rcu.so.24.0 00:02:56.743 [202/707] Generating symbol file lib/librte_meter.so.24.0.p/librte_meter.so.24.0.symbols 00:02:56.743 [203/707] Linking target lib/librte_mempool.so.24.0 00:02:56.743 [204/707] Linking target lib/librte_cfgfile.so.24.0 00:02:56.743 [205/707] Generating symbol file lib/librte_timer.so.24.0.p/librte_timer.so.24.0.symbols 00:02:56.743 [206/707] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:56.743 [207/707] Linking static target lib/librte_compressdev.a 00:02:56.743 [208/707] Generating symbol file lib/librte_rcu.so.24.0.p/librte_rcu.so.24.0.symbols 00:02:56.743 [209/707] Generating symbol file lib/librte_mempool.so.24.0.p/librte_mempool.so.24.0.symbols 00:02:56.743 [210/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:57.003 [211/707] Linking target lib/librte_mbuf.so.24.0 00:02:57.003 [212/707] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.003 [213/707] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:57.003 [214/707] Linking static target lib/librte_acl.a 00:02:57.003 [215/707] Generating symbol file lib/librte_mbuf.so.24.0.p/librte_mbuf.so.24.0.symbols 00:02:57.003 [216/707] Linking target lib/librte_net.so.24.0 00:02:57.262 [217/707] Generating symbol file lib/librte_net.so.24.0.p/librte_net.so.24.0.symbols 00:02:57.262 [218/707] Linking target lib/librte_cmdline.so.24.0 00:02:57.262 [219/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:57.262 [220/707] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.262 [221/707] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.262 [222/707] Linking target lib/librte_hash.so.24.0 00:02:57.262 [223/707] Linking target lib/librte_bbdev.so.24.0 00:02:57.262 [224/707] Linking target lib/librte_compressdev.so.24.0 00:02:57.262 [225/707] Linking target lib/librte_acl.so.24.0 00:02:57.262 [226/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:57.519 [227/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:57.519 [228/707] Generating symbol file lib/librte_hash.so.24.0.p/librte_hash.so.24.0.symbols 00:02:57.519 [229/707] Generating symbol file lib/librte_acl.so.24.0.p/librte_acl.so.24.0.symbols 00:02:57.519 [230/707] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:57.519 [231/707] Linking static target lib/librte_distributor.a 00:02:57.777 [232/707] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:02:57.777 [233/707] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:57.777 [234/707] Linking static target lib/librte_dmadev.a 00:02:57.777 [235/707] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:57.777 [236/707] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.777 [237/707] Linking target lib/librte_distributor.so.24.0 00:02:58.036 [238/707] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:58.036 [239/707] Linking target lib/librte_dmadev.so.24.0 00:02:58.295 [240/707] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:58.295 [241/707] Generating symbol file lib/librte_dmadev.so.24.0.p/librte_dmadev.so.24.0.symbols 00:02:58.295 [242/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_dma_adapter.c.o 00:02:58.295 [243/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:58.554 [244/707] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:58.554 [245/707] Linking static target lib/librte_efd.a 00:02:58.554 [246/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:58.814 [247/707] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:58.814 [248/707] Linking target lib/librte_efd.so.24.0 00:02:58.814 [249/707] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:58.814 [250/707] Linking static target lib/librte_cryptodev.a 00:02:59.074 [251/707] Compiling C object lib/librte_dispatcher.a.p/dispatcher_rte_dispatcher.c.o 00:02:59.074 [252/707] Linking static target lib/librte_dispatcher.a 00:02:59.074 [253/707] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:59.074 [254/707] Linking static target lib/librte_gpudev.a 00:02:59.074 [255/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:59.333 [256/707] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:59.333 [257/707] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:59.333 [258/707] Generating lib/dispatcher.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.591 [259/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:59.591 [260/707] Compiling C object lib/librte_gro.a.p/gro_gro_tcp6.c.o 00:02:59.850 [261/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:59.850 [262/707] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.850 [263/707] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:59.850 [264/707] Linking target lib/librte_gpudev.so.24.0 00:03:00.109 [265/707] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:03:00.109 [266/707] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:03:00.109 [267/707] Linking static target lib/librte_gro.a 00:03:00.109 [268/707] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.109 [269/707] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:03:00.109 [270/707] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:03:00.109 [271/707] Linking target lib/librte_cryptodev.so.24.0 00:03:00.109 [272/707] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.372 [273/707] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:03:00.372 [274/707] Generating symbol file lib/librte_cryptodev.so.24.0.p/librte_cryptodev.so.24.0.symbols 00:03:00.372 [275/707] Linking target lib/librte_ethdev.so.24.0 00:03:00.372 [276/707] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:03:00.372 [277/707] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:03:00.372 [278/707] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.372 [279/707] Linking static target lib/librte_eventdev.a 00:03:00.372 [280/707] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:03:00.372 [281/707] Generating symbol file lib/librte_ethdev.so.24.0.p/librte_ethdev.so.24.0.symbols 00:03:00.372 [282/707] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:03:00.372 [283/707] Linking target lib/librte_metrics.so.24.0 00:03:00.372 [284/707] Linking target lib/librte_bpf.so.24.0 00:03:00.640 [285/707] Generating symbol file lib/librte_metrics.so.24.0.p/librte_metrics.so.24.0.symbols 00:03:00.640 [286/707] Linking target lib/librte_bitratestats.so.24.0 00:03:00.640 [287/707] Generating symbol file lib/librte_bpf.so.24.0.p/librte_bpf.so.24.0.symbols 00:03:00.640 [288/707] Linking static target lib/librte_gso.a 00:03:00.640 [289/707] Linking target lib/librte_gro.so.24.0 00:03:00.640 [290/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:03:00.640 [291/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:03:00.640 [292/707] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:03:00.909 [293/707] Linking static target lib/librte_jobstats.a 00:03:00.909 [294/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:03:00.909 [295/707] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.909 [296/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:03:00.909 [297/707] Linking target lib/librte_gso.so.24.0 00:03:00.909 [298/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:03:00.909 [299/707] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:03:00.909 [300/707] Linking static target lib/librte_ip_frag.a 00:03:00.909 [301/707] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:03:01.167 [302/707] Linking target lib/librte_jobstats.so.24.0 00:03:01.167 [303/707] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:03:01.167 [304/707] Linking static target lib/librte_latencystats.a 00:03:01.167 [305/707] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:03:01.426 [306/707] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:03:01.426 [307/707] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:03:01.426 [308/707] Linking static target lib/member/libsketch_avx512_tmp.a 00:03:01.426 [309/707] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:03:01.426 [310/707] Linking target lib/librte_ip_frag.so.24.0 00:03:01.426 [311/707] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:03:01.426 [312/707] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:03:01.426 [313/707] Linking target lib/librte_latencystats.so.24.0 00:03:01.426 [314/707] Generating symbol file lib/librte_ip_frag.so.24.0.p/librte_ip_frag.so.24.0.symbols 00:03:01.426 [315/707] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:03:01.685 [316/707] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:03:01.685 [317/707] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:03:01.685 [318/707] Linking static target lib/librte_lpm.a 00:03:01.944 [319/707] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:03:01.944 [320/707] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:03:01.944 [321/707] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:03:01.944 [322/707] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:03:01.944 [323/707] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:03:01.944 [324/707] Linking static target lib/librte_pcapng.a 00:03:02.202 [325/707] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:03:02.203 [326/707] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:03:02.203 [327/707] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:03:02.203 [328/707] Linking target lib/librte_lpm.so.24.0 00:03:02.203 [329/707] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:03:02.203 [330/707] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:02.203 [331/707] Generating symbol file lib/librte_lpm.so.24.0.p/librte_lpm.so.24.0.symbols 00:03:02.203 [332/707] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:03:02.203 [333/707] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:03:02.203 [334/707] Linking target lib/librte_eventdev.so.24.0 00:03:02.203 [335/707] Linking target lib/librte_pcapng.so.24.0 00:03:02.462 [336/707] Generating symbol file lib/librte_eventdev.so.24.0.p/librte_eventdev.so.24.0.symbols 00:03:02.462 [337/707] Generating symbol file lib/librte_pcapng.so.24.0.p/librte_pcapng.so.24.0.symbols 00:03:02.462 [338/707] Linking target lib/librte_dispatcher.so.24.0 00:03:02.462 [339/707] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:03:02.722 [340/707] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:03:02.722 [341/707] Linking static target lib/librte_power.a 00:03:02.722 [342/707] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev_pmd.c.o 00:03:02.722 [343/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils.c.o 00:03:02.722 [344/707] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:03:02.722 [345/707] Linking static target lib/librte_regexdev.a 00:03:02.722 [346/707] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:03:02.722 [347/707] Linking static target lib/librte_rawdev.a 00:03:02.722 [348/707] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev.c.o 00:03:02.722 [349/707] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:03:02.980 [350/707] Linking static target lib/librte_member.a 00:03:02.980 [351/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar_bfloat16.c.o 00:03:02.980 [352/707] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar.c.o 00:03:02.980 [353/707] Linking static target lib/librte_mldev.a 00:03:02.980 [354/707] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:03:03.239 [355/707] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.239 [356/707] Linking target lib/librte_member.so.24.0 00:03:03.239 [357/707] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:03:03.239 [358/707] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.239 [359/707] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.239 [360/707] Linking target lib/librte_rawdev.so.24.0 00:03:03.239 [361/707] Linking target lib/librte_power.so.24.0 00:03:03.239 [362/707] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:03:03.239 [363/707] Linking static target lib/librte_reorder.a 00:03:03.239 [364/707] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:03:03.499 [365/707] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.499 [366/707] Linking target lib/librte_regexdev.so.24.0 00:03:03.499 [367/707] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:03:03.499 [368/707] Linking static target lib/librte_rib.a 00:03:03.499 [369/707] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:03:03.499 [370/707] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:03:03.499 [371/707] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.499 [372/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:03:03.499 [373/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:03:03.759 [374/707] Linking target lib/librte_reorder.so.24.0 00:03:03.759 [375/707] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:03:03.759 [376/707] Linking static target lib/librte_stack.a 00:03:03.759 [377/707] Generating symbol file lib/librte_reorder.so.24.0.p/librte_reorder.so.24.0.symbols 00:03:03.759 [378/707] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:03:03.759 [379/707] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:03:04.018 [380/707] Linking target lib/librte_rib.so.24.0 00:03:04.018 [381/707] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:03:04.018 [382/707] Linking static target lib/librte_security.a 00:03:04.018 [383/707] Linking target lib/librte_stack.so.24.0 00:03:04.018 [384/707] Generating symbol file lib/librte_rib.so.24.0.p/librte_rib.so.24.0.symbols 00:03:04.018 [385/707] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:03:04.018 [386/707] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:03:04.277 [387/707] Generating lib/mldev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:04.277 [388/707] Linking target lib/librte_mldev.so.24.0 00:03:04.277 [389/707] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:03:04.278 [390/707] Linking target lib/librte_security.so.24.0 00:03:04.278 [391/707] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:03:04.536 [392/707] Generating symbol file lib/librte_security.so.24.0.p/librte_security.so.24.0.symbols 00:03:04.536 [393/707] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:03:04.536 [394/707] Linking static target lib/librte_sched.a 00:03:04.536 [395/707] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:03:04.794 [396/707] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:03:04.794 [397/707] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:03:04.794 [398/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:03:04.794 [399/707] Linking target lib/librte_sched.so.24.0 00:03:05.053 [400/707] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:03:05.053 [401/707] Generating symbol file lib/librte_sched.so.24.0.p/librte_sched.so.24.0.symbols 00:03:05.053 [402/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:03:05.313 [403/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:03:05.313 [404/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:03:05.573 [405/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_ctrl_pdu.c.o 00:03:05.573 [406/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_crypto.c.o 00:03:05.573 [407/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_cnt.c.o 00:03:05.573 [408/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_reorder.c.o 00:03:05.833 [409/707] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:03:05.833 [410/707] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:03:05.833 [411/707] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:03:05.833 [412/707] Linking static target lib/librte_ipsec.a 00:03:05.833 [413/707] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:03:05.833 [414/707] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:03:06.093 [415/707] Compiling C object lib/librte_pdcp.a.p/pdcp_rte_pdcp.c.o 00:03:06.093 [416/707] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:03:06.093 [417/707] Linking target lib/librte_ipsec.so.24.0 00:03:06.353 [418/707] Generating symbol file lib/librte_ipsec.so.24.0.p/librte_ipsec.so.24.0.symbols 00:03:06.353 [419/707] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:03:06.353 [420/707] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:03:06.613 [421/707] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:03:06.614 [422/707] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:03:06.614 [423/707] Linking static target lib/librte_fib.a 00:03:06.614 [424/707] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:03:06.614 [425/707] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:03:06.873 [426/707] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:03:06.873 [427/707] Linking target lib/librte_fib.so.24.0 00:03:06.873 [428/707] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:03:06.873 [429/707] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:03:06.873 [430/707] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:03:07.133 [431/707] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_process.c.o 00:03:07.133 [432/707] Linking static target lib/librte_pdcp.a 00:03:07.393 [433/707] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:03:07.393 [434/707] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:03:07.393 [435/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:03:07.393 [436/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:03:07.653 [437/707] Generating lib/pdcp.sym_chk with a custom command (wrapped by meson to capture output) 00:03:07.653 [438/707] Linking target lib/librte_pdcp.so.24.0 00:03:07.653 [439/707] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:03:07.653 [440/707] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:03:07.914 [441/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:03:07.914 [442/707] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:03:08.174 [443/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:03:08.174 [444/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:03:08.174 [445/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:03:08.174 [446/707] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:03:08.174 [447/707] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:03:08.174 [448/707] Linking static target lib/librte_pdump.a 00:03:08.435 [449/707] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:03:08.435 [450/707] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:03:08.435 [451/707] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:03:08.435 [452/707] Linking static target lib/librte_port.a 00:03:08.435 [453/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:03:08.435 [454/707] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:03:08.435 [455/707] Linking target lib/librte_pdump.so.24.0 00:03:09.013 [456/707] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:03:09.013 [457/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:03:09.013 [458/707] Linking target lib/librte_port.so.24.0 00:03:09.013 [459/707] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:03:09.013 [460/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:03:09.013 [461/707] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:03:09.013 [462/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:03:09.013 [463/707] Generating symbol file lib/librte_port.so.24.0.p/librte_port.so.24.0.symbols 00:03:09.273 [464/707] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:03:09.273 [465/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:03:09.273 [466/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:03:09.532 [467/707] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:03:09.532 [468/707] Linking static target lib/librte_table.a 00:03:09.532 [469/707] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:03:09.792 [470/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:03:09.792 [471/707] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:03:10.051 [472/707] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:03:10.051 [473/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ipsec.c.o 00:03:10.051 [474/707] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:03:10.051 [475/707] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:03:10.051 [476/707] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:03:10.311 [477/707] Linking target lib/librte_table.so.24.0 00:03:10.311 [478/707] Generating symbol file lib/librte_table.so.24.0.p/librte_table.so.24.0.symbols 00:03:10.572 [479/707] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:03:10.572 [480/707] Compiling C object lib/librte_graph.a.p/graph_rte_graph_worker.c.o 00:03:10.572 [481/707] Compiling C object lib/librte_graph.a.p/graph_graph_pcap.c.o 00:03:10.572 [482/707] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:03:10.572 [483/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:03:10.831 [484/707] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:03:10.831 [485/707] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:03:10.831 [486/707] Compiling C object lib/librte_graph.a.p/graph_rte_graph_model_mcore_dispatch.c.o 00:03:10.831 [487/707] Linking static target lib/librte_graph.a 00:03:11.091 [488/707] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:03:11.091 [489/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:03:11.352 [490/707] Compiling C object lib/librte_node.a.p/node_ip4_local.c.o 00:03:11.352 [491/707] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:03:11.352 [492/707] Compiling C object lib/librte_node.a.p/node_ip4_reassembly.c.o 00:03:11.352 [493/707] Linking target lib/librte_graph.so.24.0 00:03:11.612 [494/707] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:03:11.612 [495/707] Generating symbol file lib/librte_graph.so.24.0.p/librte_graph.so.24.0.symbols 00:03:11.612 [496/707] Compiling C object lib/librte_node.a.p/node_null.c.o 00:03:11.612 [497/707] Compiling C object lib/librte_node.a.p/node_ip6_lookup.c.o 00:03:11.872 [498/707] Compiling C object lib/librte_node.a.p/node_kernel_tx.c.o 00:03:11.872 [499/707] Compiling C object lib/librte_node.a.p/node_log.c.o 00:03:11.872 [500/707] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:03:11.872 [501/707] Compiling C object lib/librte_node.a.p/node_ip6_rewrite.c.o 00:03:11.872 [502/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:03:12.132 [503/707] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:03:12.132 [504/707] Compiling C object lib/librte_node.a.p/node_kernel_rx.c.o 00:03:12.132 [505/707] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:03:12.392 [506/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:03:12.392 [507/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:03:12.392 [508/707] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:03:12.392 [509/707] Compiling C object lib/librte_node.a.p/node_udp4_input.c.o 00:03:12.392 [510/707] Linking static target lib/librte_node.a 00:03:12.392 [511/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:03:12.652 [512/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:03:12.652 [513/707] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:03:12.652 [514/707] Linking static target drivers/libtmp_rte_bus_pci.a 00:03:12.652 [515/707] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:03:12.652 [516/707] Linking static target drivers/libtmp_rte_bus_vdev.a 00:03:12.652 [517/707] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:03:12.652 [518/707] Linking target lib/librte_node.so.24.0 00:03:12.912 [519/707] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:03:12.912 [520/707] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:03:12.912 [521/707] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:12.912 [522/707] Linking static target drivers/librte_bus_pci.a 00:03:12.912 [523/707] Compiling C object drivers/librte_bus_pci.so.24.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:12.912 [524/707] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:12.913 [525/707] Linking static target drivers/librte_bus_vdev.a 00:03:13.173 [526/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:03:13.173 [527/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:03:13.173 [528/707] Compiling C object drivers/librte_bus_vdev.so.24.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:13.173 [529/707] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:13.173 [530/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:03:13.173 [531/707] Linking target drivers/librte_bus_vdev.so.24.0 00:03:13.173 [532/707] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:03:13.173 [533/707] Linking static target drivers/libtmp_rte_mempool_ring.a 00:03:13.433 [534/707] Generating symbol file drivers/librte_bus_vdev.so.24.0.p/librte_bus_vdev.so.24.0.symbols 00:03:13.433 [535/707] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:03:13.433 [536/707] Linking target drivers/librte_bus_pci.so.24.0 00:03:13.433 [537/707] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:03:13.433 [538/707] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:13.433 [539/707] Linking static target drivers/librte_mempool_ring.a 00:03:13.433 [540/707] Generating symbol file drivers/librte_bus_pci.so.24.0.p/librte_bus_pci.so.24.0.symbols 00:03:13.433 [541/707] Compiling C object drivers/librte_mempool_ring.so.24.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:13.433 [542/707] Linking target drivers/librte_mempool_ring.so.24.0 00:03:13.433 [543/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:03:13.693 [544/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:03:13.952 [545/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:03:14.211 [546/707] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:03:14.211 [547/707] Linking static target drivers/net/i40e/base/libi40e_base.a 00:03:14.470 [548/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:03:15.040 [549/707] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:03:15.040 [550/707] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:03:15.040 [551/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:03:15.040 [552/707] Compiling C object drivers/net/i40e/libi40e_avx2_lib.a.p/i40e_rxtx_vec_avx2.c.o 00:03:15.040 [553/707] Linking static target drivers/net/i40e/libi40e_avx2_lib.a 00:03:15.300 [554/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:03:15.300 [555/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:03:15.560 [556/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:03:15.560 [557/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:03:15.820 [558/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:03:15.820 [559/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_recycle_mbufs_vec_common.c.o 00:03:16.080 [560/707] Compiling C object app/dpdk-graph.p/graph_cli.c.o 00:03:16.339 [561/707] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:03:16.339 [562/707] Compiling C object app/dpdk-graph.p/graph_conn.c.o 00:03:16.339 [563/707] Compiling C object app/dpdk-graph.p/graph_ethdev_rx.c.o 00:03:16.599 [564/707] Compiling C object app/dpdk-graph.p/graph_ethdev.c.o 00:03:16.599 [565/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:03:16.599 [566/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:03:16.599 [567/707] Compiling C object app/dpdk-graph.p/graph_ip4_route.c.o 00:03:16.599 [568/707] Compiling C object app/dpdk-graph.p/graph_ip6_route.c.o 00:03:16.859 [569/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:03:16.859 [570/707] Compiling C object app/dpdk-graph.p/graph_graph.c.o 00:03:16.859 [571/707] Compiling C object app/dpdk-graph.p/graph_main.c.o 00:03:16.859 [572/707] Compiling C object app/dpdk-graph.p/graph_l3fwd.c.o 00:03:16.859 [573/707] Compiling C object app/dpdk-graph.p/graph_mempool.c.o 00:03:17.118 [574/707] Compiling C object app/dpdk-graph.p/graph_utils.c.o 00:03:17.118 [575/707] Compiling C object app/dpdk-graph.p/graph_neigh.c.o 00:03:17.376 [576/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:03:17.377 [577/707] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:03:17.377 [578/707] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:03:17.377 [579/707] Linking static target drivers/libtmp_rte_net_i40e.a 00:03:17.377 [580/707] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:03:17.377 [581/707] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:03:17.694 [582/707] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:03:17.694 [583/707] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:17.694 [584/707] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:03:17.694 [585/707] Compiling C object drivers/librte_net_i40e.so.24.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:17.694 [586/707] Linking static target drivers/librte_net_i40e.a 00:03:17.954 [587/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:03:17.954 [588/707] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:03:17.954 [589/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:03:18.214 [590/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:03:18.214 [591/707] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:03:18.214 [592/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:03:18.474 [593/707] Linking target drivers/librte_net_i40e.so.24.0 00:03:18.474 [594/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:03:18.474 [595/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:03:18.733 [596/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:03:18.733 [597/707] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:03:18.733 [598/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:03:18.993 [599/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:03:18.993 [600/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:03:19.252 [601/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:03:19.252 [602/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:03:19.252 [603/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:03:19.252 [604/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:03:19.252 [605/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:03:19.252 [606/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:03:19.511 [607/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:03:19.511 [608/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:03:19.511 [609/707] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:03:19.511 [610/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:03:19.771 [611/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:03:19.771 [612/707] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_main.c.o 00:03:19.771 [613/707] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_benchmark.c.o 00:03:20.031 [614/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:03:20.031 [615/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:03:20.290 [616/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:03:20.859 [617/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:03:20.859 [618/707] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:03:20.859 [619/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:03:20.859 [620/707] Linking static target lib/librte_vhost.a 00:03:20.859 [621/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:03:20.859 [622/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:03:21.118 [623/707] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:03:21.118 [624/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:03:21.118 [625/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:03:21.377 [626/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:03:21.377 [627/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:03:21.377 [628/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_test.c.o 00:03:21.377 [629/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_main.c.o 00:03:21.377 [630/707] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:03:21.377 [631/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_parser.c.o 00:03:21.636 [632/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_options.c.o 00:03:21.636 [633/707] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:03:21.895 [634/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_device_ops.c.o 00:03:21.895 [635/707] Linking target lib/librte_vhost.so.24.0 00:03:21.895 [636/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_common.c.o 00:03:21.895 [637/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_common.c.o 00:03:21.895 [638/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_ops.c.o 00:03:22.154 [639/707] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:03:22.154 [640/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_ordered.c.o 00:03:22.154 [641/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_interleave.c.o 00:03:22.154 [642/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_stats.c.o 00:03:22.413 [643/707] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:03:22.413 [644/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:03:22.413 [645/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:03:22.413 [646/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:03:22.672 [647/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:03:22.672 [648/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:03:22.672 [649/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:03:22.672 [650/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:03:22.930 [651/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:03:22.930 [652/707] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:03:23.189 [653/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_cman.c.o 00:03:23.189 [654/707] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:03:23.448 [655/707] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_common.c.o 00:03:23.448 [656/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:03:23.448 [657/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:03:23.448 [658/707] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:03:23.448 [659/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:03:23.708 [660/707] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:03:23.967 [661/707] Linking static target lib/librte_pipeline.a 00:03:23.967 [662/707] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:03:23.967 [663/707] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:03:23.967 [664/707] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:03:23.967 [665/707] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:03:23.967 [666/707] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:03:24.227 [667/707] Linking target app/dpdk-dumpcap 00:03:24.227 [668/707] Linking target app/dpdk-graph 00:03:24.227 [669/707] Linking target app/dpdk-pdump 00:03:24.227 [670/707] Linking target app/dpdk-test-acl 00:03:24.487 [671/707] Linking target app/dpdk-proc-info 00:03:24.487 [672/707] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:03:24.487 [673/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:03:24.487 [674/707] Linking target app/dpdk-test-bbdev 00:03:24.748 [675/707] Linking target app/dpdk-test-cmdline 00:03:24.748 [676/707] Linking target app/dpdk-test-compress-perf 00:03:24.748 [677/707] Linking target app/dpdk-test-crypto-perf 00:03:24.748 [678/707] Linking target app/dpdk-test-dma-perf 00:03:24.748 [679/707] Linking target app/dpdk-test-eventdev 00:03:25.008 [680/707] Linking target app/dpdk-test-fib 00:03:25.008 [681/707] Linking target app/dpdk-test-flow-perf 00:03:25.008 [682/707] Linking target app/dpdk-test-gpudev 00:03:25.008 [683/707] Linking target app/dpdk-test-mldev 00:03:25.008 [684/707] Linking target app/dpdk-test-pipeline 00:03:25.268 [685/707] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:03:25.526 [686/707] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:03:25.526 [687/707] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:03:25.526 [688/707] Compiling C object app/dpdk-testpmd.p/test-pmd_recycle_mbufs.c.o 00:03:25.526 [689/707] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:03:25.785 [690/707] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:03:25.785 [691/707] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:03:26.044 [692/707] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:03:26.044 [693/707] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:03:26.044 [694/707] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:26.044 [695/707] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:03:26.044 [696/707] Linking target lib/librte_pipeline.so.24.0 00:03:26.305 [697/707] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:03:26.305 [698/707] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:03:26.305 [699/707] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:03:26.305 [700/707] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:03:26.563 [701/707] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:03:26.563 [702/707] Linking target app/dpdk-test-regex 00:03:26.563 [703/707] Linking target app/dpdk-test-sad 00:03:26.823 [704/707] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:03:26.823 [705/707] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:03:27.083 [706/707] Linking target app/dpdk-test-security-perf 00:03:27.342 [707/707] Linking target app/dpdk-testpmd 00:03:27.343 10:52:32 build_native_dpdk -- common/autobuild_common.sh@194 -- $ uname -s 00:03:27.343 10:52:32 build_native_dpdk -- common/autobuild_common.sh@194 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:03:27.343 10:52:32 build_native_dpdk -- common/autobuild_common.sh@207 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:03:27.343 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:27.343 [0/1] Installing files. 00:03:27.605 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:03:27.605 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:27.605 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:27.605 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.606 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-macsec/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:27.607 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.608 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ipsec_sa.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/rss.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/efd_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/commands.list to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.609 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:27.610 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:27.610 Installing lib/librte_log.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.610 Installing lib/librte_log.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_kvargs.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_telemetry.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_eal.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_rcu.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_mempool.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_mbuf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_net.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_meter.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ethdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cmdline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_metrics.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_hash.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_timer.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_acl.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bbdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bitratestats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_bpf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cfgfile.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_compressdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_cryptodev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_distributor.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_dmadev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_efd.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_eventdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_dispatcher.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_dispatcher.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gpudev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gro.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_gso.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_ip_frag.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.870 Installing lib/librte_jobstats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_latencystats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_lpm.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_member.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pcapng.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_power.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_rawdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_regexdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_mldev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_mldev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_rib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_reorder.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_sched.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_security.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_stack.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_vhost.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_ipsec.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pdcp.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pdcp.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_fib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_port.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pdump.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_table.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_pipeline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:27.871 Installing lib/librte_graph.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing lib/librte_node.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing drivers/librte_bus_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:28.134 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing drivers/librte_bus_vdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:28.134 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing drivers/librte_mempool_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:28.134 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.134 Installing drivers/librte_net_i40e.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0 00:03:28.134 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-graph to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-dma-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-mldev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/log/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lock_annotations.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.134 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_stdatomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_dtls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_pdcp_hdr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.135 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_dma_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/dispatcher/rte_dispatcher.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/mldev/rte_mldev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/pdcp/rte_pdcp_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.136 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_mcore_dispatch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_model_rtc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip6_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_udp4_input_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/buildtools/dpdk-cmdline-gen.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-rss-flows.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:28.137 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:28.137 Installing symlink pointing to librte_log.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so.24 00:03:28.137 Installing symlink pointing to librte_log.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_log.so 00:03:28.137 Installing symlink pointing to librte_kvargs.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.24 00:03:28.137 Installing symlink pointing to librte_kvargs.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:03:28.137 Installing symlink pointing to librte_telemetry.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.24 00:03:28.137 Installing symlink pointing to librte_telemetry.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:03:28.137 Installing symlink pointing to librte_eal.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.24 00:03:28.137 Installing symlink pointing to librte_eal.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:03:28.137 Installing symlink pointing to librte_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.24 00:03:28.137 Installing symlink pointing to librte_ring.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:03:28.137 Installing symlink pointing to librte_rcu.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.24 00:03:28.137 Installing symlink pointing to librte_rcu.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:03:28.137 Installing symlink pointing to librte_mempool.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.24 00:03:28.137 Installing symlink pointing to librte_mempool.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:03:28.137 Installing symlink pointing to librte_mbuf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.24 00:03:28.137 Installing symlink pointing to librte_mbuf.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:03:28.137 Installing symlink pointing to librte_net.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.24 00:03:28.137 Installing symlink pointing to librte_net.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:03:28.137 Installing symlink pointing to librte_meter.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.24 00:03:28.137 Installing symlink pointing to librte_meter.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:03:28.137 Installing symlink pointing to librte_ethdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.24 00:03:28.137 Installing symlink pointing to librte_ethdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:03:28.137 Installing symlink pointing to librte_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.24 00:03:28.137 Installing symlink pointing to librte_pci.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:03:28.137 Installing symlink pointing to librte_cmdline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.24 00:03:28.137 Installing symlink pointing to librte_cmdline.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:03:28.137 Installing symlink pointing to librte_metrics.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.24 00:03:28.137 Installing symlink pointing to librte_metrics.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:03:28.137 Installing symlink pointing to librte_hash.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.24 00:03:28.137 Installing symlink pointing to librte_hash.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:03:28.137 Installing symlink pointing to librte_timer.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.24 00:03:28.137 Installing symlink pointing to librte_timer.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:03:28.137 Installing symlink pointing to librte_acl.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.24 00:03:28.137 Installing symlink pointing to librte_acl.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:03:28.137 Installing symlink pointing to librte_bbdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.24 00:03:28.137 Installing symlink pointing to librte_bbdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:03:28.137 Installing symlink pointing to librte_bitratestats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.24 00:03:28.137 Installing symlink pointing to librte_bitratestats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:03:28.137 Installing symlink pointing to librte_bpf.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.24 00:03:28.137 Installing symlink pointing to librte_bpf.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:03:28.137 Installing symlink pointing to librte_cfgfile.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.24 00:03:28.137 Installing symlink pointing to librte_cfgfile.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:03:28.137 Installing symlink pointing to librte_compressdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.24 00:03:28.137 Installing symlink pointing to librte_compressdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:03:28.137 Installing symlink pointing to librte_cryptodev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.24 00:03:28.137 Installing symlink pointing to librte_cryptodev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:03:28.137 Installing symlink pointing to librte_distributor.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.24 00:03:28.137 Installing symlink pointing to librte_distributor.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:03:28.137 Installing symlink pointing to librte_dmadev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.24 00:03:28.137 Installing symlink pointing to librte_dmadev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:03:28.137 Installing symlink pointing to librte_efd.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.24 00:03:28.137 Installing symlink pointing to librte_efd.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:03:28.137 Installing symlink pointing to librte_eventdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.24 00:03:28.137 Installing symlink pointing to librte_eventdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:03:28.137 Installing symlink pointing to librte_dispatcher.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so.24 00:03:28.137 Installing symlink pointing to librte_dispatcher.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dispatcher.so 00:03:28.137 Installing symlink pointing to librte_gpudev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.24 00:03:28.137 Installing symlink pointing to librte_gpudev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:03:28.137 Installing symlink pointing to librte_gro.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.24 00:03:28.137 Installing symlink pointing to librte_gro.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:03:28.137 Installing symlink pointing to librte_gso.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.24 00:03:28.137 Installing symlink pointing to librte_gso.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:03:28.137 Installing symlink pointing to librte_ip_frag.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.24 00:03:28.137 Installing symlink pointing to librte_ip_frag.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:03:28.137 Installing symlink pointing to librte_jobstats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.24 00:03:28.137 Installing symlink pointing to librte_jobstats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:03:28.137 Installing symlink pointing to librte_latencystats.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.24 00:03:28.137 Installing symlink pointing to librte_latencystats.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:03:28.137 Installing symlink pointing to librte_lpm.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.24 00:03:28.137 Installing symlink pointing to librte_lpm.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:03:28.137 Installing symlink pointing to librte_member.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.24 00:03:28.137 Installing symlink pointing to librte_member.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:03:28.137 Installing symlink pointing to librte_pcapng.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.24 00:03:28.137 Installing symlink pointing to librte_pcapng.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:03:28.137 Installing symlink pointing to librte_power.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.24 00:03:28.137 Installing symlink pointing to librte_power.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:03:28.138 Installing symlink pointing to librte_rawdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.24 00:03:28.138 Installing symlink pointing to librte_rawdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:03:28.138 Installing symlink pointing to librte_regexdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.24 00:03:28.138 Installing symlink pointing to librte_regexdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:03:28.138 Installing symlink pointing to librte_mldev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so.24 00:03:28.138 Installing symlink pointing to librte_mldev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mldev.so 00:03:28.138 Installing symlink pointing to librte_rib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.24 00:03:28.138 Installing symlink pointing to librte_rib.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:03:28.138 Installing symlink pointing to librte_reorder.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.24 00:03:28.138 Installing symlink pointing to librte_reorder.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:03:28.138 Installing symlink pointing to librte_sched.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.24 00:03:28.138 Installing symlink pointing to librte_sched.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:03:28.138 Installing symlink pointing to librte_security.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.24 00:03:28.138 Installing symlink pointing to librte_security.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:03:28.138 './librte_bus_pci.so' -> 'dpdk/pmds-24.0/librte_bus_pci.so' 00:03:28.138 './librte_bus_pci.so.24' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24' 00:03:28.138 './librte_bus_pci.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24.0' 00:03:28.138 './librte_bus_vdev.so' -> 'dpdk/pmds-24.0/librte_bus_vdev.so' 00:03:28.138 './librte_bus_vdev.so.24' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24' 00:03:28.138 './librte_bus_vdev.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24.0' 00:03:28.138 './librte_mempool_ring.so' -> 'dpdk/pmds-24.0/librte_mempool_ring.so' 00:03:28.138 './librte_mempool_ring.so.24' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24' 00:03:28.138 './librte_mempool_ring.so.24.0' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24.0' 00:03:28.138 './librte_net_i40e.so' -> 'dpdk/pmds-24.0/librte_net_i40e.so' 00:03:28.138 './librte_net_i40e.so.24' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24' 00:03:28.138 './librte_net_i40e.so.24.0' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24.0' 00:03:28.138 Installing symlink pointing to librte_stack.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.24 00:03:28.138 Installing symlink pointing to librte_stack.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:03:28.138 Installing symlink pointing to librte_vhost.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.24 00:03:28.138 Installing symlink pointing to librte_vhost.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:03:28.138 Installing symlink pointing to librte_ipsec.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.24 00:03:28.138 Installing symlink pointing to librte_ipsec.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:03:28.138 Installing symlink pointing to librte_pdcp.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so.24 00:03:28.138 Installing symlink pointing to librte_pdcp.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdcp.so 00:03:28.138 Installing symlink pointing to librte_fib.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.24 00:03:28.138 Installing symlink pointing to librte_fib.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:03:28.138 Installing symlink pointing to librte_port.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.24 00:03:28.138 Installing symlink pointing to librte_port.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:03:28.138 Installing symlink pointing to librte_pdump.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.24 00:03:28.138 Installing symlink pointing to librte_pdump.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:03:28.138 Installing symlink pointing to librte_table.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.24 00:03:28.138 Installing symlink pointing to librte_table.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:03:28.138 Installing symlink pointing to librte_pipeline.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.24 00:03:28.138 Installing symlink pointing to librte_pipeline.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:03:28.138 Installing symlink pointing to librte_graph.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.24 00:03:28.138 Installing symlink pointing to librte_graph.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:03:28.138 Installing symlink pointing to librte_node.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.24 00:03:28.138 Installing symlink pointing to librte_node.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:03:28.138 Installing symlink pointing to librte_bus_pci.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24 00:03:28.138 Installing symlink pointing to librte_bus_pci.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:03:28.138 Installing symlink pointing to librte_bus_vdev.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24 00:03:28.138 Installing symlink pointing to librte_bus_vdev.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:03:28.138 Installing symlink pointing to librte_mempool_ring.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24 00:03:28.138 Installing symlink pointing to librte_mempool_ring.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:03:28.138 Installing symlink pointing to librte_net_i40e.so.24.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24 00:03:28.138 Installing symlink pointing to librte_net_i40e.so.24 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:03:28.138 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-24.0' 00:03:28.138 10:52:33 build_native_dpdk -- common/autobuild_common.sh@213 -- $ cat 00:03:28.138 10:52:33 build_native_dpdk -- common/autobuild_common.sh@218 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:28.138 00:03:28.138 real 0m49.546s 00:03:28.138 user 5m33.879s 00:03:28.138 sys 0m58.786s 00:03:28.138 10:52:33 build_native_dpdk -- common/autotest_common.sh@1128 -- $ xtrace_disable 00:03:28.138 10:52:33 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:03:28.138 ************************************ 00:03:28.138 END TEST build_native_dpdk 00:03:28.138 ************************************ 00:03:28.138 10:52:33 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:03:28.138 10:52:33 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:03:28.138 10:52:33 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:03:28.397 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:03:28.656 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:03:28.656 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:03:28.656 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:03:28.942 Using 'verbs' RDMA provider 00:03:45.214 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:04:00.110 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:04:00.370 Creating mk/config.mk...done. 00:04:00.370 Creating mk/cc.flags.mk...done. 00:04:00.370 Type 'make' to build. 00:04:00.370 10:53:05 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:04:00.370 10:53:05 -- common/autotest_common.sh@1103 -- $ '[' 3 -le 1 ']' 00:04:00.370 10:53:05 -- common/autotest_common.sh@1109 -- $ xtrace_disable 00:04:00.370 10:53:05 -- common/autotest_common.sh@10 -- $ set +x 00:04:00.370 ************************************ 00:04:00.370 START TEST make 00:04:00.370 ************************************ 00:04:00.370 10:53:05 make -- common/autotest_common.sh@1127 -- $ make -j10 00:04:00.940 make[1]: Nothing to be done for 'all'. 00:04:47.655 CC lib/log/log.o 00:04:47.655 CC lib/log/log_flags.o 00:04:47.655 CC lib/log/log_deprecated.o 00:04:47.655 CC lib/ut_mock/mock.o 00:04:47.655 CC lib/ut/ut.o 00:04:47.655 LIB libspdk_log.a 00:04:47.655 LIB libspdk_ut.a 00:04:47.655 LIB libspdk_ut_mock.a 00:04:47.655 SO libspdk_log.so.7.1 00:04:47.655 SO libspdk_ut.so.2.0 00:04:47.655 SO libspdk_ut_mock.so.6.0 00:04:47.655 SYMLINK libspdk_log.so 00:04:47.655 SYMLINK libspdk_ut_mock.so 00:04:47.655 SYMLINK libspdk_ut.so 00:04:47.655 CC lib/util/base64.o 00:04:47.655 CC lib/util/bit_array.o 00:04:47.655 CC lib/util/crc16.o 00:04:47.655 CC lib/util/crc32.o 00:04:47.655 CC lib/util/crc32c.o 00:04:47.655 CC lib/util/cpuset.o 00:04:47.655 CC lib/dma/dma.o 00:04:47.655 CC lib/ioat/ioat.o 00:04:47.655 CXX lib/trace_parser/trace.o 00:04:47.655 CC lib/vfio_user/host/vfio_user_pci.o 00:04:47.655 CC lib/util/crc32_ieee.o 00:04:47.655 CC lib/vfio_user/host/vfio_user.o 00:04:47.655 CC lib/util/crc64.o 00:04:47.655 CC lib/util/dif.o 00:04:47.655 LIB libspdk_dma.a 00:04:47.655 CC lib/util/fd.o 00:04:47.655 SO libspdk_dma.so.5.0 00:04:47.655 CC lib/util/fd_group.o 00:04:47.655 CC lib/util/file.o 00:04:47.655 CC lib/util/hexlify.o 00:04:47.655 SYMLINK libspdk_dma.so 00:04:47.655 CC lib/util/iov.o 00:04:47.655 LIB libspdk_ioat.a 00:04:47.655 CC lib/util/math.o 00:04:47.655 SO libspdk_ioat.so.7.0 00:04:47.655 CC lib/util/net.o 00:04:47.655 LIB libspdk_vfio_user.a 00:04:47.655 CC lib/util/pipe.o 00:04:47.655 SYMLINK libspdk_ioat.so 00:04:47.655 SO libspdk_vfio_user.so.5.0 00:04:47.656 CC lib/util/strerror_tls.o 00:04:47.656 CC lib/util/string.o 00:04:47.656 SYMLINK libspdk_vfio_user.so 00:04:47.656 CC lib/util/uuid.o 00:04:47.656 CC lib/util/xor.o 00:04:47.656 CC lib/util/zipf.o 00:04:47.656 CC lib/util/md5.o 00:04:47.656 LIB libspdk_util.a 00:04:47.656 SO libspdk_util.so.10.0 00:04:47.656 LIB libspdk_trace_parser.a 00:04:47.656 SO libspdk_trace_parser.so.6.0 00:04:47.656 SYMLINK libspdk_util.so 00:04:47.656 SYMLINK libspdk_trace_parser.so 00:04:47.656 CC lib/env_dpdk/env.o 00:04:47.656 CC lib/env_dpdk/memory.o 00:04:47.656 CC lib/env_dpdk/pci.o 00:04:47.656 CC lib/env_dpdk/init.o 00:04:47.656 CC lib/rdma_provider/common.o 00:04:47.656 CC lib/rdma_utils/rdma_utils.o 00:04:47.656 CC lib/json/json_parse.o 00:04:47.656 CC lib/vmd/vmd.o 00:04:47.656 CC lib/conf/conf.o 00:04:47.656 CC lib/idxd/idxd.o 00:04:47.656 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:47.656 LIB libspdk_conf.a 00:04:47.656 CC lib/json/json_util.o 00:04:47.656 SO libspdk_conf.so.6.0 00:04:47.656 LIB libspdk_rdma_utils.a 00:04:47.656 SO libspdk_rdma_utils.so.1.0 00:04:47.656 SYMLINK libspdk_conf.so 00:04:47.656 CC lib/json/json_write.o 00:04:47.656 SYMLINK libspdk_rdma_utils.so 00:04:47.656 CC lib/vmd/led.o 00:04:47.656 CC lib/idxd/idxd_user.o 00:04:47.656 CC lib/env_dpdk/threads.o 00:04:47.656 LIB libspdk_rdma_provider.a 00:04:47.656 SO libspdk_rdma_provider.so.6.0 00:04:47.656 SYMLINK libspdk_rdma_provider.so 00:04:47.656 CC lib/idxd/idxd_kernel.o 00:04:47.656 CC lib/env_dpdk/pci_ioat.o 00:04:47.656 CC lib/env_dpdk/pci_virtio.o 00:04:47.656 CC lib/env_dpdk/pci_vmd.o 00:04:47.656 CC lib/env_dpdk/pci_idxd.o 00:04:47.656 LIB libspdk_json.a 00:04:47.656 CC lib/env_dpdk/pci_event.o 00:04:47.656 CC lib/env_dpdk/sigbus_handler.o 00:04:47.656 CC lib/env_dpdk/pci_dpdk.o 00:04:47.656 SO libspdk_json.so.6.0 00:04:47.656 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:47.656 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:47.656 LIB libspdk_idxd.a 00:04:47.656 SYMLINK libspdk_json.so 00:04:47.656 LIB libspdk_vmd.a 00:04:47.656 SO libspdk_idxd.so.12.1 00:04:47.656 SO libspdk_vmd.so.6.0 00:04:47.656 SYMLINK libspdk_idxd.so 00:04:47.656 SYMLINK libspdk_vmd.so 00:04:47.915 CC lib/jsonrpc/jsonrpc_server.o 00:04:47.915 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:47.915 CC lib/jsonrpc/jsonrpc_client.o 00:04:47.915 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:48.174 LIB libspdk_jsonrpc.a 00:04:48.174 SO libspdk_jsonrpc.so.6.0 00:04:48.433 SYMLINK libspdk_jsonrpc.so 00:04:48.691 LIB libspdk_env_dpdk.a 00:04:48.691 CC lib/rpc/rpc.o 00:04:48.691 SO libspdk_env_dpdk.so.15.1 00:04:48.949 SYMLINK libspdk_env_dpdk.so 00:04:48.949 LIB libspdk_rpc.a 00:04:48.949 SO libspdk_rpc.so.6.0 00:04:49.207 SYMLINK libspdk_rpc.so 00:04:49.466 CC lib/trace/trace.o 00:04:49.466 CC lib/trace/trace_flags.o 00:04:49.466 CC lib/trace/trace_rpc.o 00:04:49.466 CC lib/notify/notify.o 00:04:49.466 CC lib/notify/notify_rpc.o 00:04:49.466 CC lib/keyring/keyring_rpc.o 00:04:49.466 CC lib/keyring/keyring.o 00:04:49.724 LIB libspdk_notify.a 00:04:49.724 SO libspdk_notify.so.6.0 00:04:49.724 LIB libspdk_keyring.a 00:04:49.724 LIB libspdk_trace.a 00:04:49.724 SO libspdk_keyring.so.2.0 00:04:49.724 SYMLINK libspdk_notify.so 00:04:49.724 SO libspdk_trace.so.11.0 00:04:49.724 SYMLINK libspdk_keyring.so 00:04:49.724 SYMLINK libspdk_trace.so 00:04:50.308 CC lib/thread/thread.o 00:04:50.308 CC lib/thread/iobuf.o 00:04:50.308 CC lib/sock/sock.o 00:04:50.308 CC lib/sock/sock_rpc.o 00:04:50.566 LIB libspdk_sock.a 00:04:50.824 SO libspdk_sock.so.10.0 00:04:50.824 SYMLINK libspdk_sock.so 00:04:51.083 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:51.083 CC lib/nvme/nvme_ctrlr.o 00:04:51.083 CC lib/nvme/nvme_fabric.o 00:04:51.083 CC lib/nvme/nvme.o 00:04:51.083 CC lib/nvme/nvme_pcie_common.o 00:04:51.083 CC lib/nvme/nvme_pcie.o 00:04:51.083 CC lib/nvme/nvme_ns_cmd.o 00:04:51.083 CC lib/nvme/nvme_ns.o 00:04:51.083 CC lib/nvme/nvme_qpair.o 00:04:52.022 CC lib/nvme/nvme_quirks.o 00:04:52.022 CC lib/nvme/nvme_transport.o 00:04:52.022 LIB libspdk_thread.a 00:04:52.022 CC lib/nvme/nvme_discovery.o 00:04:52.022 SO libspdk_thread.so.11.0 00:04:52.022 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:52.022 SYMLINK libspdk_thread.so 00:04:52.022 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:52.022 CC lib/nvme/nvme_tcp.o 00:04:52.022 CC lib/nvme/nvme_opal.o 00:04:52.022 CC lib/accel/accel.o 00:04:52.282 CC lib/accel/accel_rpc.o 00:04:52.282 CC lib/nvme/nvme_io_msg.o 00:04:52.541 CC lib/nvme/nvme_poll_group.o 00:04:52.541 CC lib/accel/accel_sw.o 00:04:52.541 CC lib/nvme/nvme_zns.o 00:04:52.801 CC lib/blob/blobstore.o 00:04:52.801 CC lib/blob/request.o 00:04:52.801 CC lib/nvme/nvme_stubs.o 00:04:52.801 CC lib/nvme/nvme_auth.o 00:04:53.060 CC lib/blob/zeroes.o 00:04:53.060 CC lib/blob/blob_bs_dev.o 00:04:53.060 CC lib/nvme/nvme_cuse.o 00:04:53.319 CC lib/nvme/nvme_rdma.o 00:04:53.319 CC lib/init/json_config.o 00:04:53.319 CC lib/virtio/virtio.o 00:04:53.319 CC lib/virtio/virtio_vhost_user.o 00:04:53.319 LIB libspdk_accel.a 00:04:53.319 SO libspdk_accel.so.16.0 00:04:53.578 CC lib/init/subsystem.o 00:04:53.578 SYMLINK libspdk_accel.so 00:04:53.578 CC lib/init/subsystem_rpc.o 00:04:53.578 CC lib/virtio/virtio_vfio_user.o 00:04:53.578 CC lib/init/rpc.o 00:04:53.578 CC lib/virtio/virtio_pci.o 00:04:53.836 CC lib/fsdev/fsdev.o 00:04:53.836 LIB libspdk_init.a 00:04:53.836 CC lib/fsdev/fsdev_io.o 00:04:53.836 CC lib/fsdev/fsdev_rpc.o 00:04:53.836 SO libspdk_init.so.6.0 00:04:53.836 CC lib/bdev/bdev.o 00:04:53.836 CC lib/bdev/bdev_rpc.o 00:04:54.095 SYMLINK libspdk_init.so 00:04:54.095 CC lib/bdev/bdev_zone.o 00:04:54.095 LIB libspdk_virtio.a 00:04:54.095 CC lib/bdev/part.o 00:04:54.095 SO libspdk_virtio.so.7.0 00:04:54.095 SYMLINK libspdk_virtio.so 00:04:54.095 CC lib/bdev/scsi_nvme.o 00:04:54.352 CC lib/event/app.o 00:04:54.352 CC lib/event/reactor.o 00:04:54.352 CC lib/event/log_rpc.o 00:04:54.352 CC lib/event/app_rpc.o 00:04:54.352 CC lib/event/scheduler_static.o 00:04:54.609 LIB libspdk_fsdev.a 00:04:54.609 SO libspdk_fsdev.so.2.0 00:04:54.609 SYMLINK libspdk_fsdev.so 00:04:54.868 LIB libspdk_event.a 00:04:54.868 SO libspdk_event.so.14.0 00:04:54.868 SYMLINK libspdk_event.so 00:04:54.868 LIB libspdk_nvme.a 00:04:55.127 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:55.127 SO libspdk_nvme.so.14.1 00:04:55.385 SYMLINK libspdk_nvme.so 00:04:55.952 LIB libspdk_fuse_dispatcher.a 00:04:55.952 SO libspdk_fuse_dispatcher.so.1.0 00:04:55.952 SYMLINK libspdk_fuse_dispatcher.so 00:04:56.518 LIB libspdk_blob.a 00:04:56.518 SO libspdk_blob.so.11.0 00:04:56.776 SYMLINK libspdk_blob.so 00:04:57.036 LIB libspdk_bdev.a 00:04:57.036 CC lib/lvol/lvol.o 00:04:57.036 CC lib/blobfs/blobfs.o 00:04:57.036 CC lib/blobfs/tree.o 00:04:57.036 SO libspdk_bdev.so.17.0 00:04:57.296 SYMLINK libspdk_bdev.so 00:04:57.555 CC lib/nbd/nbd.o 00:04:57.555 CC lib/nvmf/ctrlr.o 00:04:57.555 CC lib/nbd/nbd_rpc.o 00:04:57.555 CC lib/nvmf/ctrlr_discovery.o 00:04:57.555 CC lib/nvmf/ctrlr_bdev.o 00:04:57.555 CC lib/ftl/ftl_core.o 00:04:57.555 CC lib/ublk/ublk.o 00:04:57.555 CC lib/scsi/dev.o 00:04:57.815 CC lib/ublk/ublk_rpc.o 00:04:57.815 CC lib/scsi/lun.o 00:04:57.815 CC lib/nvmf/subsystem.o 00:04:57.815 CC lib/ftl/ftl_init.o 00:04:58.075 LIB libspdk_nbd.a 00:04:58.075 SO libspdk_nbd.so.7.0 00:04:58.075 LIB libspdk_blobfs.a 00:04:58.075 SO libspdk_blobfs.so.10.0 00:04:58.075 CC lib/ftl/ftl_layout.o 00:04:58.075 SYMLINK libspdk_nbd.so 00:04:58.075 CC lib/scsi/port.o 00:04:58.075 CC lib/ftl/ftl_debug.o 00:04:58.075 SYMLINK libspdk_blobfs.so 00:04:58.075 CC lib/ftl/ftl_io.o 00:04:58.075 CC lib/ftl/ftl_sb.o 00:04:58.075 LIB libspdk_lvol.a 00:04:58.335 SO libspdk_lvol.so.10.0 00:04:58.335 CC lib/scsi/scsi.o 00:04:58.335 LIB libspdk_ublk.a 00:04:58.335 SO libspdk_ublk.so.3.0 00:04:58.335 SYMLINK libspdk_lvol.so 00:04:58.335 CC lib/scsi/scsi_bdev.o 00:04:58.335 CC lib/scsi/scsi_pr.o 00:04:58.335 SYMLINK libspdk_ublk.so 00:04:58.335 CC lib/ftl/ftl_l2p.o 00:04:58.335 CC lib/ftl/ftl_l2p_flat.o 00:04:58.335 CC lib/nvmf/nvmf.o 00:04:58.335 CC lib/nvmf/nvmf_rpc.o 00:04:58.335 CC lib/nvmf/transport.o 00:04:58.335 CC lib/ftl/ftl_nv_cache.o 00:04:58.594 CC lib/nvmf/tcp.o 00:04:58.594 CC lib/scsi/scsi_rpc.o 00:04:58.594 CC lib/ftl/ftl_band.o 00:04:58.594 CC lib/ftl/ftl_band_ops.o 00:04:58.853 CC lib/scsi/task.o 00:04:59.112 LIB libspdk_scsi.a 00:04:59.112 CC lib/ftl/ftl_writer.o 00:04:59.112 CC lib/ftl/ftl_rq.o 00:04:59.112 SO libspdk_scsi.so.9.0 00:04:59.112 SYMLINK libspdk_scsi.so 00:04:59.112 CC lib/ftl/ftl_reloc.o 00:04:59.112 CC lib/ftl/ftl_l2p_cache.o 00:04:59.371 CC lib/nvmf/stubs.o 00:04:59.371 CC lib/nvmf/mdns_server.o 00:04:59.371 CC lib/nvmf/rdma.o 00:04:59.371 CC lib/nvmf/auth.o 00:04:59.371 CC lib/iscsi/conn.o 00:04:59.629 CC lib/iscsi/init_grp.o 00:04:59.629 CC lib/iscsi/iscsi.o 00:04:59.629 CC lib/ftl/ftl_p2l.o 00:04:59.629 CC lib/ftl/ftl_p2l_log.o 00:04:59.889 CC lib/iscsi/param.o 00:04:59.889 CC lib/ftl/mngt/ftl_mngt.o 00:04:59.889 CC lib/vhost/vhost.o 00:05:00.149 CC lib/vhost/vhost_rpc.o 00:05:00.149 CC lib/vhost/vhost_scsi.o 00:05:00.149 CC lib/vhost/vhost_blk.o 00:05:00.149 CC lib/iscsi/portal_grp.o 00:05:00.149 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:05:00.409 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:05:00.409 CC lib/vhost/rte_vhost_user.o 00:05:00.409 CC lib/ftl/mngt/ftl_mngt_startup.o 00:05:00.409 CC lib/ftl/mngt/ftl_mngt_md.o 00:05:00.669 CC lib/ftl/mngt/ftl_mngt_misc.o 00:05:00.669 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:05:00.669 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:05:00.669 CC lib/ftl/mngt/ftl_mngt_band.o 00:05:00.669 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:05:00.928 CC lib/iscsi/tgt_node.o 00:05:00.928 CC lib/iscsi/iscsi_subsystem.o 00:05:00.928 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:05:00.928 CC lib/iscsi/iscsi_rpc.o 00:05:00.928 CC lib/iscsi/task.o 00:05:00.928 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:05:01.188 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:05:01.188 CC lib/ftl/utils/ftl_conf.o 00:05:01.188 CC lib/ftl/utils/ftl_md.o 00:05:01.188 CC lib/ftl/utils/ftl_mempool.o 00:05:01.188 CC lib/ftl/utils/ftl_bitmap.o 00:05:01.447 CC lib/ftl/utils/ftl_property.o 00:05:01.447 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:05:01.447 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:05:01.447 LIB libspdk_iscsi.a 00:05:01.447 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:05:01.447 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:05:01.447 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:05:01.447 SO libspdk_iscsi.so.8.0 00:05:01.705 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:05:01.705 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:05:01.705 LIB libspdk_vhost.a 00:05:01.705 CC lib/ftl/upgrade/ftl_sb_v3.o 00:05:01.705 CC lib/ftl/upgrade/ftl_sb_v5.o 00:05:01.705 CC lib/ftl/nvc/ftl_nvc_dev.o 00:05:01.705 SO libspdk_vhost.so.8.0 00:05:01.705 SYMLINK libspdk_iscsi.so 00:05:01.705 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:05:01.705 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:05:01.705 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:05:01.705 SYMLINK libspdk_vhost.so 00:05:01.705 CC lib/ftl/base/ftl_base_dev.o 00:05:01.705 CC lib/ftl/base/ftl_base_bdev.o 00:05:01.964 CC lib/ftl/ftl_trace.o 00:05:01.964 LIB libspdk_nvmf.a 00:05:01.964 SO libspdk_nvmf.so.20.0 00:05:02.224 LIB libspdk_ftl.a 00:05:02.224 SYMLINK libspdk_nvmf.so 00:05:02.484 SO libspdk_ftl.so.9.0 00:05:02.745 SYMLINK libspdk_ftl.so 00:05:03.004 CC module/env_dpdk/env_dpdk_rpc.o 00:05:03.004 CC module/accel/dsa/accel_dsa.o 00:05:03.004 CC module/keyring/file/keyring.o 00:05:03.004 CC module/blob/bdev/blob_bdev.o 00:05:03.004 CC module/keyring/linux/keyring.o 00:05:03.004 CC module/scheduler/dynamic/scheduler_dynamic.o 00:05:03.004 CC module/accel/ioat/accel_ioat.o 00:05:03.004 CC module/sock/posix/posix.o 00:05:03.004 CC module/fsdev/aio/fsdev_aio.o 00:05:03.265 CC module/accel/error/accel_error.o 00:05:03.265 LIB libspdk_env_dpdk_rpc.a 00:05:03.265 SO libspdk_env_dpdk_rpc.so.6.0 00:05:03.265 CC module/keyring/linux/keyring_rpc.o 00:05:03.265 SYMLINK libspdk_env_dpdk_rpc.so 00:05:03.265 CC module/keyring/file/keyring_rpc.o 00:05:03.265 CC module/accel/error/accel_error_rpc.o 00:05:03.265 CC module/accel/ioat/accel_ioat_rpc.o 00:05:03.265 LIB libspdk_scheduler_dynamic.a 00:05:03.265 SO libspdk_scheduler_dynamic.so.4.0 00:05:03.265 CC module/fsdev/aio/fsdev_aio_rpc.o 00:05:03.265 LIB libspdk_keyring_linux.a 00:05:03.265 SYMLINK libspdk_scheduler_dynamic.so 00:05:03.265 LIB libspdk_keyring_file.a 00:05:03.265 LIB libspdk_accel_error.a 00:05:03.525 LIB libspdk_blob_bdev.a 00:05:03.525 SO libspdk_keyring_linux.so.1.0 00:05:03.525 CC module/accel/dsa/accel_dsa_rpc.o 00:05:03.525 SO libspdk_keyring_file.so.2.0 00:05:03.525 SO libspdk_accel_error.so.2.0 00:05:03.525 SO libspdk_blob_bdev.so.11.0 00:05:03.525 LIB libspdk_accel_ioat.a 00:05:03.525 SYMLINK libspdk_keyring_linux.so 00:05:03.525 SYMLINK libspdk_blob_bdev.so 00:05:03.525 SYMLINK libspdk_keyring_file.so 00:05:03.525 SO libspdk_accel_ioat.so.6.0 00:05:03.525 CC module/fsdev/aio/linux_aio_mgr.o 00:05:03.525 SYMLINK libspdk_accel_error.so 00:05:03.525 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:05:03.525 LIB libspdk_accel_dsa.a 00:05:03.525 SYMLINK libspdk_accel_ioat.so 00:05:03.525 SO libspdk_accel_dsa.so.5.0 00:05:03.785 CC module/scheduler/gscheduler/gscheduler.o 00:05:03.785 SYMLINK libspdk_accel_dsa.so 00:05:03.785 CC module/accel/iaa/accel_iaa.o 00:05:03.785 CC module/accel/iaa/accel_iaa_rpc.o 00:05:03.785 LIB libspdk_scheduler_dpdk_governor.a 00:05:03.785 SO libspdk_scheduler_dpdk_governor.so.4.0 00:05:03.785 CC module/bdev/error/vbdev_error.o 00:05:03.785 CC module/bdev/delay/vbdev_delay.o 00:05:03.785 CC module/blobfs/bdev/blobfs_bdev.o 00:05:03.785 LIB libspdk_scheduler_gscheduler.a 00:05:03.785 CC module/bdev/gpt/gpt.o 00:05:03.785 SYMLINK libspdk_scheduler_dpdk_governor.so 00:05:03.785 CC module/bdev/delay/vbdev_delay_rpc.o 00:05:03.785 CC module/bdev/error/vbdev_error_rpc.o 00:05:03.785 SO libspdk_scheduler_gscheduler.so.4.0 00:05:03.785 LIB libspdk_fsdev_aio.a 00:05:03.785 LIB libspdk_accel_iaa.a 00:05:03.785 SYMLINK libspdk_scheduler_gscheduler.so 00:05:03.785 SO libspdk_fsdev_aio.so.1.0 00:05:03.785 SO libspdk_accel_iaa.so.3.0 00:05:04.044 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:05:04.044 LIB libspdk_sock_posix.a 00:05:04.044 SYMLINK libspdk_accel_iaa.so 00:05:04.044 SYMLINK libspdk_fsdev_aio.so 00:05:04.044 SO libspdk_sock_posix.so.6.0 00:05:04.044 CC module/bdev/gpt/vbdev_gpt.o 00:05:04.044 LIB libspdk_bdev_error.a 00:05:04.044 SYMLINK libspdk_sock_posix.so 00:05:04.044 SO libspdk_bdev_error.so.6.0 00:05:04.044 LIB libspdk_blobfs_bdev.a 00:05:04.044 SO libspdk_blobfs_bdev.so.6.0 00:05:04.044 CC module/bdev/malloc/bdev_malloc.o 00:05:04.044 CC module/bdev/lvol/vbdev_lvol.o 00:05:04.044 SYMLINK libspdk_bdev_error.so 00:05:04.044 CC module/bdev/null/bdev_null.o 00:05:04.044 CC module/bdev/nvme/bdev_nvme.o 00:05:04.304 LIB libspdk_bdev_delay.a 00:05:04.304 SYMLINK libspdk_blobfs_bdev.so 00:05:04.304 CC module/bdev/passthru/vbdev_passthru.o 00:05:04.304 SO libspdk_bdev_delay.so.6.0 00:05:04.304 CC module/bdev/raid/bdev_raid.o 00:05:04.304 SYMLINK libspdk_bdev_delay.so 00:05:04.304 CC module/bdev/malloc/bdev_malloc_rpc.o 00:05:04.304 CC module/bdev/split/vbdev_split.o 00:05:04.304 LIB libspdk_bdev_gpt.a 00:05:04.304 CC module/bdev/zone_block/vbdev_zone_block.o 00:05:04.304 SO libspdk_bdev_gpt.so.6.0 00:05:04.304 SYMLINK libspdk_bdev_gpt.so 00:05:04.304 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:05:04.562 CC module/bdev/null/bdev_null_rpc.o 00:05:04.562 CC module/bdev/split/vbdev_split_rpc.o 00:05:04.562 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:05:04.562 CC module/bdev/raid/bdev_raid_rpc.o 00:05:04.562 LIB libspdk_bdev_malloc.a 00:05:04.562 SO libspdk_bdev_malloc.so.6.0 00:05:04.562 LIB libspdk_bdev_null.a 00:05:04.562 LIB libspdk_bdev_split.a 00:05:04.562 SO libspdk_bdev_null.so.6.0 00:05:04.562 SYMLINK libspdk_bdev_malloc.so 00:05:04.562 SO libspdk_bdev_split.so.6.0 00:05:04.562 CC module/bdev/nvme/bdev_nvme_rpc.o 00:05:04.562 LIB libspdk_bdev_passthru.a 00:05:04.562 SYMLINK libspdk_bdev_null.so 00:05:04.562 CC module/bdev/nvme/nvme_rpc.o 00:05:04.820 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:05:04.820 SO libspdk_bdev_passthru.so.6.0 00:05:04.820 SYMLINK libspdk_bdev_split.so 00:05:04.820 CC module/bdev/raid/bdev_raid_sb.o 00:05:04.820 CC module/bdev/nvme/bdev_mdns_client.o 00:05:04.820 SYMLINK libspdk_bdev_passthru.so 00:05:04.820 LIB libspdk_bdev_lvol.a 00:05:04.820 CC module/bdev/aio/bdev_aio.o 00:05:04.820 LIB libspdk_bdev_zone_block.a 00:05:04.820 SO libspdk_bdev_lvol.so.6.0 00:05:04.820 SO libspdk_bdev_zone_block.so.6.0 00:05:05.079 CC module/bdev/nvme/vbdev_opal.o 00:05:05.079 SYMLINK libspdk_bdev_lvol.so 00:05:05.079 CC module/bdev/nvme/vbdev_opal_rpc.o 00:05:05.079 CC module/bdev/ftl/bdev_ftl.o 00:05:05.079 SYMLINK libspdk_bdev_zone_block.so 00:05:05.079 CC module/bdev/ftl/bdev_ftl_rpc.o 00:05:05.079 CC module/bdev/raid/raid0.o 00:05:05.079 CC module/bdev/iscsi/bdev_iscsi.o 00:05:05.079 CC module/bdev/aio/bdev_aio_rpc.o 00:05:05.079 CC module/bdev/raid/raid1.o 00:05:05.339 CC module/bdev/raid/concat.o 00:05:05.339 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:05:05.339 LIB libspdk_bdev_ftl.a 00:05:05.339 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:05:05.339 SO libspdk_bdev_ftl.so.6.0 00:05:05.339 LIB libspdk_bdev_aio.a 00:05:05.339 CC module/bdev/raid/raid5f.o 00:05:05.339 SYMLINK libspdk_bdev_ftl.so 00:05:05.339 SO libspdk_bdev_aio.so.6.0 00:05:05.339 SYMLINK libspdk_bdev_aio.so 00:05:05.598 LIB libspdk_bdev_iscsi.a 00:05:05.598 SO libspdk_bdev_iscsi.so.6.0 00:05:05.598 CC module/bdev/virtio/bdev_virtio_scsi.o 00:05:05.598 CC module/bdev/virtio/bdev_virtio_blk.o 00:05:05.598 CC module/bdev/virtio/bdev_virtio_rpc.o 00:05:05.598 SYMLINK libspdk_bdev_iscsi.so 00:05:05.858 LIB libspdk_bdev_raid.a 00:05:05.858 SO libspdk_bdev_raid.so.6.0 00:05:06.117 SYMLINK libspdk_bdev_raid.so 00:05:06.117 LIB libspdk_bdev_virtio.a 00:05:06.117 SO libspdk_bdev_virtio.so.6.0 00:05:06.117 SYMLINK libspdk_bdev_virtio.so 00:05:07.497 LIB libspdk_bdev_nvme.a 00:05:07.497 SO libspdk_bdev_nvme.so.7.1 00:05:07.497 SYMLINK libspdk_bdev_nvme.so 00:05:08.104 CC module/event/subsystems/iobuf/iobuf.o 00:05:08.104 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:05:08.104 CC module/event/subsystems/vmd/vmd.o 00:05:08.104 CC module/event/subsystems/vmd/vmd_rpc.o 00:05:08.104 CC module/event/subsystems/sock/sock.o 00:05:08.104 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:05:08.104 CC module/event/subsystems/fsdev/fsdev.o 00:05:08.104 CC module/event/subsystems/keyring/keyring.o 00:05:08.104 CC module/event/subsystems/scheduler/scheduler.o 00:05:08.365 LIB libspdk_event_vmd.a 00:05:08.365 LIB libspdk_event_iobuf.a 00:05:08.365 LIB libspdk_event_keyring.a 00:05:08.365 LIB libspdk_event_fsdev.a 00:05:08.365 LIB libspdk_event_sock.a 00:05:08.365 LIB libspdk_event_vhost_blk.a 00:05:08.365 LIB libspdk_event_scheduler.a 00:05:08.365 SO libspdk_event_keyring.so.1.0 00:05:08.365 SO libspdk_event_fsdev.so.1.0 00:05:08.365 SO libspdk_event_vmd.so.6.0 00:05:08.365 SO libspdk_event_sock.so.5.0 00:05:08.365 SO libspdk_event_iobuf.so.3.0 00:05:08.365 SO libspdk_event_vhost_blk.so.3.0 00:05:08.365 SO libspdk_event_scheduler.so.4.0 00:05:08.365 SYMLINK libspdk_event_keyring.so 00:05:08.365 SYMLINK libspdk_event_fsdev.so 00:05:08.365 SYMLINK libspdk_event_sock.so 00:05:08.365 SYMLINK libspdk_event_iobuf.so 00:05:08.365 SYMLINK libspdk_event_vhost_blk.so 00:05:08.365 SYMLINK libspdk_event_vmd.so 00:05:08.365 SYMLINK libspdk_event_scheduler.so 00:05:08.625 CC module/event/subsystems/accel/accel.o 00:05:08.883 LIB libspdk_event_accel.a 00:05:08.883 SO libspdk_event_accel.so.6.0 00:05:08.883 SYMLINK libspdk_event_accel.so 00:05:09.452 CC module/event/subsystems/bdev/bdev.o 00:05:09.712 LIB libspdk_event_bdev.a 00:05:09.712 SO libspdk_event_bdev.so.6.0 00:05:09.712 SYMLINK libspdk_event_bdev.so 00:05:09.971 CC module/event/subsystems/scsi/scsi.o 00:05:09.971 CC module/event/subsystems/nbd/nbd.o 00:05:09.971 CC module/event/subsystems/ublk/ublk.o 00:05:09.971 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:05:09.971 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:05:10.231 LIB libspdk_event_nbd.a 00:05:10.231 LIB libspdk_event_scsi.a 00:05:10.231 SO libspdk_event_nbd.so.6.0 00:05:10.231 LIB libspdk_event_ublk.a 00:05:10.231 SO libspdk_event_scsi.so.6.0 00:05:10.231 SO libspdk_event_ublk.so.3.0 00:05:10.231 SYMLINK libspdk_event_nbd.so 00:05:10.231 SYMLINK libspdk_event_scsi.so 00:05:10.231 SYMLINK libspdk_event_ublk.so 00:05:10.231 LIB libspdk_event_nvmf.a 00:05:10.490 SO libspdk_event_nvmf.so.6.0 00:05:10.490 SYMLINK libspdk_event_nvmf.so 00:05:10.749 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:05:10.749 CC module/event/subsystems/iscsi/iscsi.o 00:05:10.749 LIB libspdk_event_vhost_scsi.a 00:05:10.749 LIB libspdk_event_iscsi.a 00:05:10.749 SO libspdk_event_vhost_scsi.so.3.0 00:05:11.009 SO libspdk_event_iscsi.so.6.0 00:05:11.009 SYMLINK libspdk_event_vhost_scsi.so 00:05:11.009 SYMLINK libspdk_event_iscsi.so 00:05:11.269 SO libspdk.so.6.0 00:05:11.269 SYMLINK libspdk.so 00:05:11.528 CC app/trace_record/trace_record.o 00:05:11.528 CXX app/trace/trace.o 00:05:11.528 CC examples/interrupt_tgt/interrupt_tgt.o 00:05:11.528 CC app/nvmf_tgt/nvmf_main.o 00:05:11.528 CC app/iscsi_tgt/iscsi_tgt.o 00:05:11.528 CC examples/util/zipf/zipf.o 00:05:11.528 CC examples/ioat/perf/perf.o 00:05:11.528 CC test/thread/poller_perf/poller_perf.o 00:05:11.528 CC test/app/bdev_svc/bdev_svc.o 00:05:11.528 CC test/dma/test_dma/test_dma.o 00:05:11.787 LINK interrupt_tgt 00:05:11.787 LINK nvmf_tgt 00:05:11.787 LINK zipf 00:05:11.787 LINK poller_perf 00:05:11.787 LINK iscsi_tgt 00:05:11.787 LINK spdk_trace_record 00:05:11.787 LINK bdev_svc 00:05:11.787 LINK ioat_perf 00:05:12.046 LINK spdk_trace 00:05:12.046 CC app/spdk_lspci/spdk_lspci.o 00:05:12.046 CC app/spdk_nvme_perf/perf.o 00:05:12.047 CC examples/ioat/verify/verify.o 00:05:12.047 CC test/app/histogram_perf/histogram_perf.o 00:05:12.047 CC app/spdk_tgt/spdk_tgt.o 00:05:12.047 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:05:12.047 CC test/app/jsoncat/jsoncat.o 00:05:12.306 LINK spdk_lspci 00:05:12.306 LINK test_dma 00:05:12.306 CC test/app/stub/stub.o 00:05:12.306 CC examples/thread/thread/thread_ex.o 00:05:12.306 LINK histogram_perf 00:05:12.306 LINK jsoncat 00:05:12.306 LINK verify 00:05:12.306 LINK spdk_tgt 00:05:12.306 LINK stub 00:05:12.565 CC app/spdk_nvme_identify/identify.o 00:05:12.565 CC app/spdk_nvme_discover/discovery_aer.o 00:05:12.565 CC app/spdk_top/spdk_top.o 00:05:12.565 LINK thread 00:05:12.565 TEST_HEADER include/spdk/accel.h 00:05:12.565 TEST_HEADER include/spdk/accel_module.h 00:05:12.565 TEST_HEADER include/spdk/assert.h 00:05:12.565 TEST_HEADER include/spdk/barrier.h 00:05:12.565 TEST_HEADER include/spdk/base64.h 00:05:12.565 TEST_HEADER include/spdk/bdev.h 00:05:12.565 TEST_HEADER include/spdk/bdev_module.h 00:05:12.565 TEST_HEADER include/spdk/bdev_zone.h 00:05:12.565 TEST_HEADER include/spdk/bit_array.h 00:05:12.565 TEST_HEADER include/spdk/bit_pool.h 00:05:12.565 TEST_HEADER include/spdk/blob_bdev.h 00:05:12.565 TEST_HEADER include/spdk/blobfs_bdev.h 00:05:12.565 TEST_HEADER include/spdk/blobfs.h 00:05:12.565 TEST_HEADER include/spdk/blob.h 00:05:12.565 TEST_HEADER include/spdk/conf.h 00:05:12.565 TEST_HEADER include/spdk/config.h 00:05:12.565 TEST_HEADER include/spdk/cpuset.h 00:05:12.565 TEST_HEADER include/spdk/crc16.h 00:05:12.565 TEST_HEADER include/spdk/crc32.h 00:05:12.565 TEST_HEADER include/spdk/crc64.h 00:05:12.565 TEST_HEADER include/spdk/dif.h 00:05:12.565 TEST_HEADER include/spdk/dma.h 00:05:12.565 TEST_HEADER include/spdk/endian.h 00:05:12.565 TEST_HEADER include/spdk/env_dpdk.h 00:05:12.565 TEST_HEADER include/spdk/env.h 00:05:12.565 TEST_HEADER include/spdk/event.h 00:05:12.565 TEST_HEADER include/spdk/fd_group.h 00:05:12.565 TEST_HEADER include/spdk/fd.h 00:05:12.565 TEST_HEADER include/spdk/file.h 00:05:12.565 TEST_HEADER include/spdk/fsdev.h 00:05:12.565 TEST_HEADER include/spdk/fsdev_module.h 00:05:12.565 TEST_HEADER include/spdk/ftl.h 00:05:12.565 TEST_HEADER include/spdk/fuse_dispatcher.h 00:05:12.565 TEST_HEADER include/spdk/gpt_spec.h 00:05:12.565 TEST_HEADER include/spdk/hexlify.h 00:05:12.565 TEST_HEADER include/spdk/histogram_data.h 00:05:12.565 TEST_HEADER include/spdk/idxd.h 00:05:12.565 LINK nvme_fuzz 00:05:12.565 TEST_HEADER include/spdk/idxd_spec.h 00:05:12.565 TEST_HEADER include/spdk/init.h 00:05:12.565 TEST_HEADER include/spdk/ioat.h 00:05:12.565 TEST_HEADER include/spdk/ioat_spec.h 00:05:12.565 TEST_HEADER include/spdk/iscsi_spec.h 00:05:12.565 TEST_HEADER include/spdk/json.h 00:05:12.565 TEST_HEADER include/spdk/jsonrpc.h 00:05:12.565 TEST_HEADER include/spdk/keyring.h 00:05:12.565 TEST_HEADER include/spdk/keyring_module.h 00:05:12.565 TEST_HEADER include/spdk/likely.h 00:05:12.565 TEST_HEADER include/spdk/log.h 00:05:12.565 TEST_HEADER include/spdk/lvol.h 00:05:12.565 TEST_HEADER include/spdk/md5.h 00:05:12.565 TEST_HEADER include/spdk/memory.h 00:05:12.565 TEST_HEADER include/spdk/mmio.h 00:05:12.565 TEST_HEADER include/spdk/nbd.h 00:05:12.565 TEST_HEADER include/spdk/net.h 00:05:12.565 TEST_HEADER include/spdk/notify.h 00:05:12.565 TEST_HEADER include/spdk/nvme.h 00:05:12.565 TEST_HEADER include/spdk/nvme_intel.h 00:05:12.565 TEST_HEADER include/spdk/nvme_ocssd.h 00:05:12.565 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:05:12.565 TEST_HEADER include/spdk/nvme_spec.h 00:05:12.565 CC examples/sock/hello_world/hello_sock.o 00:05:12.565 TEST_HEADER include/spdk/nvme_zns.h 00:05:12.824 TEST_HEADER include/spdk/nvmf_cmd.h 00:05:12.824 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:05:12.824 TEST_HEADER include/spdk/nvmf.h 00:05:12.824 TEST_HEADER include/spdk/nvmf_spec.h 00:05:12.824 TEST_HEADER include/spdk/nvmf_transport.h 00:05:12.824 TEST_HEADER include/spdk/opal.h 00:05:12.824 TEST_HEADER include/spdk/opal_spec.h 00:05:12.824 TEST_HEADER include/spdk/pci_ids.h 00:05:12.824 TEST_HEADER include/spdk/pipe.h 00:05:12.824 TEST_HEADER include/spdk/queue.h 00:05:12.824 TEST_HEADER include/spdk/reduce.h 00:05:12.824 TEST_HEADER include/spdk/rpc.h 00:05:12.824 TEST_HEADER include/spdk/scheduler.h 00:05:12.824 TEST_HEADER include/spdk/scsi.h 00:05:12.824 TEST_HEADER include/spdk/scsi_spec.h 00:05:12.824 TEST_HEADER include/spdk/sock.h 00:05:12.824 TEST_HEADER include/spdk/stdinc.h 00:05:12.824 TEST_HEADER include/spdk/string.h 00:05:12.824 TEST_HEADER include/spdk/thread.h 00:05:12.824 TEST_HEADER include/spdk/trace.h 00:05:12.824 LINK spdk_nvme_discover 00:05:12.824 TEST_HEADER include/spdk/trace_parser.h 00:05:12.824 TEST_HEADER include/spdk/tree.h 00:05:12.824 TEST_HEADER include/spdk/ublk.h 00:05:12.824 CC app/vhost/vhost.o 00:05:12.824 TEST_HEADER include/spdk/util.h 00:05:12.824 TEST_HEADER include/spdk/uuid.h 00:05:12.824 TEST_HEADER include/spdk/version.h 00:05:12.824 TEST_HEADER include/spdk/vfio_user_pci.h 00:05:12.824 TEST_HEADER include/spdk/vfio_user_spec.h 00:05:12.824 TEST_HEADER include/spdk/vhost.h 00:05:12.824 TEST_HEADER include/spdk/vmd.h 00:05:12.824 CC examples/vmd/lsvmd/lsvmd.o 00:05:12.824 TEST_HEADER include/spdk/xor.h 00:05:12.824 TEST_HEADER include/spdk/zipf.h 00:05:12.824 CXX test/cpp_headers/accel.o 00:05:12.824 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:05:13.084 LINK lsvmd 00:05:13.084 CC examples/idxd/perf/perf.o 00:05:13.084 LINK hello_sock 00:05:13.084 CXX test/cpp_headers/accel_module.o 00:05:13.084 LINK vhost 00:05:13.084 CXX test/cpp_headers/assert.o 00:05:13.084 LINK spdk_nvme_perf 00:05:13.084 CXX test/cpp_headers/barrier.o 00:05:13.084 CXX test/cpp_headers/base64.o 00:05:13.084 CC examples/vmd/led/led.o 00:05:13.344 CXX test/cpp_headers/bdev.o 00:05:13.344 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:05:13.344 CC test/env/vtophys/vtophys.o 00:05:13.344 LINK idxd_perf 00:05:13.344 LINK led 00:05:13.344 CC test/env/mem_callbacks/mem_callbacks.o 00:05:13.344 CC app/spdk_dd/spdk_dd.o 00:05:13.344 LINK spdk_nvme_identify 00:05:13.605 LINK vtophys 00:05:13.605 CXX test/cpp_headers/bdev_module.o 00:05:13.605 LINK env_dpdk_post_init 00:05:13.605 LINK spdk_top 00:05:13.605 CC app/fio/nvme/fio_plugin.o 00:05:13.605 CXX test/cpp_headers/bdev_zone.o 00:05:13.605 CC examples/fsdev/hello_world/hello_fsdev.o 00:05:13.865 CC app/fio/bdev/fio_plugin.o 00:05:13.865 CC test/env/memory/memory_ut.o 00:05:13.865 LINK spdk_dd 00:05:13.865 CC test/env/pci/pci_ut.o 00:05:13.865 CC test/event/event_perf/event_perf.o 00:05:13.865 CXX test/cpp_headers/bit_array.o 00:05:13.865 LINK mem_callbacks 00:05:13.865 LINK hello_fsdev 00:05:14.123 LINK event_perf 00:05:14.123 CXX test/cpp_headers/bit_pool.o 00:05:14.123 CXX test/cpp_headers/blob_bdev.o 00:05:14.123 CXX test/cpp_headers/blobfs_bdev.o 00:05:14.382 CC test/nvme/aer/aer.o 00:05:14.382 CC test/event/reactor/reactor.o 00:05:14.382 CC test/event/reactor_perf/reactor_perf.o 00:05:14.382 LINK pci_ut 00:05:14.382 LINK spdk_nvme 00:05:14.382 LINK spdk_bdev 00:05:14.382 CC examples/accel/perf/accel_perf.o 00:05:14.382 CXX test/cpp_headers/blobfs.o 00:05:14.382 LINK reactor_perf 00:05:14.382 LINK reactor 00:05:14.382 CC test/event/app_repeat/app_repeat.o 00:05:14.382 CC test/rpc_client/rpc_client_test.o 00:05:14.640 CXX test/cpp_headers/blob.o 00:05:14.640 LINK aer 00:05:14.641 LINK app_repeat 00:05:14.641 CXX test/cpp_headers/conf.o 00:05:14.641 CC test/event/scheduler/scheduler.o 00:05:14.641 LINK rpc_client_test 00:05:14.641 CC test/accel/dif/dif.o 00:05:14.900 CC test/blobfs/mkfs/mkfs.o 00:05:14.900 CC test/nvme/reset/reset.o 00:05:14.900 CXX test/cpp_headers/config.o 00:05:14.900 CXX test/cpp_headers/cpuset.o 00:05:14.900 LINK iscsi_fuzz 00:05:14.900 LINK accel_perf 00:05:14.900 LINK scheduler 00:05:14.900 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:05:14.900 LINK memory_ut 00:05:14.900 LINK mkfs 00:05:14.900 CXX test/cpp_headers/crc16.o 00:05:15.158 CC test/lvol/esnap/esnap.o 00:05:15.158 LINK reset 00:05:15.158 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:05:15.158 CXX test/cpp_headers/crc32.o 00:05:15.158 CXX test/cpp_headers/crc64.o 00:05:15.158 CXX test/cpp_headers/dif.o 00:05:15.158 CC test/nvme/sgl/sgl.o 00:05:15.158 CXX test/cpp_headers/dma.o 00:05:15.417 CC examples/blob/hello_world/hello_blob.o 00:05:15.417 CC examples/nvme/hello_world/hello_world.o 00:05:15.417 CXX test/cpp_headers/endian.o 00:05:15.417 CC examples/bdev/hello_world/hello_bdev.o 00:05:15.417 CC examples/nvme/reconnect/reconnect.o 00:05:15.417 CC examples/nvme/nvme_manage/nvme_manage.o 00:05:15.417 LINK dif 00:05:15.417 LINK vhost_fuzz 00:05:15.417 LINK sgl 00:05:15.417 CXX test/cpp_headers/env_dpdk.o 00:05:15.417 LINK hello_world 00:05:15.682 LINK hello_blob 00:05:15.682 LINK hello_bdev 00:05:15.682 CXX test/cpp_headers/env.o 00:05:15.682 CXX test/cpp_headers/event.o 00:05:15.682 CC examples/nvme/arbitration/arbitration.o 00:05:15.682 LINK reconnect 00:05:15.682 CC examples/nvme/hotplug/hotplug.o 00:05:15.682 CC test/nvme/e2edp/nvme_dp.o 00:05:15.949 CXX test/cpp_headers/fd_group.o 00:05:15.949 CC examples/bdev/bdevperf/bdevperf.o 00:05:15.949 CC examples/blob/cli/blobcli.o 00:05:15.949 LINK nvme_manage 00:05:15.949 CC examples/nvme/cmb_copy/cmb_copy.o 00:05:15.949 LINK hotplug 00:05:15.949 CXX test/cpp_headers/fd.o 00:05:15.949 CC test/bdev/bdevio/bdevio.o 00:05:15.949 LINK arbitration 00:05:15.949 LINK nvme_dp 00:05:16.208 LINK cmb_copy 00:05:16.208 CXX test/cpp_headers/file.o 00:05:16.208 CC examples/nvme/abort/abort.o 00:05:16.208 CXX test/cpp_headers/fsdev.o 00:05:16.208 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:05:16.208 CXX test/cpp_headers/fsdev_module.o 00:05:16.208 CC test/nvme/overhead/overhead.o 00:05:16.209 CC test/nvme/err_injection/err_injection.o 00:05:16.209 CXX test/cpp_headers/ftl.o 00:05:16.469 LINK blobcli 00:05:16.469 LINK bdevio 00:05:16.469 LINK pmr_persistence 00:05:16.469 CXX test/cpp_headers/fuse_dispatcher.o 00:05:16.469 LINK abort 00:05:16.469 LINK err_injection 00:05:16.469 CXX test/cpp_headers/gpt_spec.o 00:05:16.469 CXX test/cpp_headers/hexlify.o 00:05:16.469 CXX test/cpp_headers/histogram_data.o 00:05:16.469 LINK overhead 00:05:16.728 CC test/nvme/startup/startup.o 00:05:16.728 CC test/nvme/reserve/reserve.o 00:05:16.728 LINK bdevperf 00:05:16.728 CXX test/cpp_headers/idxd.o 00:05:16.728 CXX test/cpp_headers/idxd_spec.o 00:05:16.728 CC test/nvme/connect_stress/connect_stress.o 00:05:16.728 CC test/nvme/boot_partition/boot_partition.o 00:05:16.728 CC test/nvme/simple_copy/simple_copy.o 00:05:16.728 LINK startup 00:05:16.728 CC test/nvme/compliance/nvme_compliance.o 00:05:16.728 LINK reserve 00:05:16.728 CXX test/cpp_headers/init.o 00:05:16.728 CXX test/cpp_headers/ioat.o 00:05:16.988 LINK boot_partition 00:05:16.988 LINK connect_stress 00:05:16.988 LINK simple_copy 00:05:16.988 CC test/nvme/fused_ordering/fused_ordering.o 00:05:16.988 CXX test/cpp_headers/ioat_spec.o 00:05:16.988 CC test/nvme/doorbell_aers/doorbell_aers.o 00:05:16.988 CXX test/cpp_headers/iscsi_spec.o 00:05:16.988 CC examples/nvmf/nvmf/nvmf.o 00:05:16.988 CC test/nvme/fdp/fdp.o 00:05:16.988 CC test/nvme/cuse/cuse.o 00:05:16.988 CXX test/cpp_headers/json.o 00:05:16.988 LINK nvme_compliance 00:05:17.248 CXX test/cpp_headers/jsonrpc.o 00:05:17.248 LINK fused_ordering 00:05:17.248 CXX test/cpp_headers/keyring.o 00:05:17.248 LINK doorbell_aers 00:05:17.248 CXX test/cpp_headers/keyring_module.o 00:05:17.248 CXX test/cpp_headers/likely.o 00:05:17.248 CXX test/cpp_headers/log.o 00:05:17.248 CXX test/cpp_headers/lvol.o 00:05:17.248 CXX test/cpp_headers/md5.o 00:05:17.248 LINK nvmf 00:05:17.248 CXX test/cpp_headers/memory.o 00:05:17.248 CXX test/cpp_headers/mmio.o 00:05:17.248 LINK fdp 00:05:17.248 CXX test/cpp_headers/nbd.o 00:05:17.508 CXX test/cpp_headers/net.o 00:05:17.508 CXX test/cpp_headers/notify.o 00:05:17.508 CXX test/cpp_headers/nvme.o 00:05:17.508 CXX test/cpp_headers/nvme_intel.o 00:05:17.508 CXX test/cpp_headers/nvme_ocssd.o 00:05:17.508 CXX test/cpp_headers/nvme_ocssd_spec.o 00:05:17.508 CXX test/cpp_headers/nvme_spec.o 00:05:17.508 CXX test/cpp_headers/nvme_zns.o 00:05:17.508 CXX test/cpp_headers/nvmf_cmd.o 00:05:17.508 CXX test/cpp_headers/nvmf_fc_spec.o 00:05:17.508 CXX test/cpp_headers/nvmf.o 00:05:17.508 CXX test/cpp_headers/nvmf_spec.o 00:05:17.768 CXX test/cpp_headers/nvmf_transport.o 00:05:17.768 CXX test/cpp_headers/opal.o 00:05:17.768 CXX test/cpp_headers/opal_spec.o 00:05:17.768 CXX test/cpp_headers/pci_ids.o 00:05:17.768 CXX test/cpp_headers/pipe.o 00:05:17.768 CXX test/cpp_headers/queue.o 00:05:17.768 CXX test/cpp_headers/reduce.o 00:05:17.768 CXX test/cpp_headers/rpc.o 00:05:17.768 CXX test/cpp_headers/scheduler.o 00:05:17.768 CXX test/cpp_headers/scsi.o 00:05:17.768 CXX test/cpp_headers/scsi_spec.o 00:05:17.768 CXX test/cpp_headers/sock.o 00:05:17.768 CXX test/cpp_headers/stdinc.o 00:05:17.768 CXX test/cpp_headers/string.o 00:05:17.768 CXX test/cpp_headers/thread.o 00:05:18.028 CXX test/cpp_headers/trace.o 00:05:18.028 CXX test/cpp_headers/trace_parser.o 00:05:18.028 CXX test/cpp_headers/tree.o 00:05:18.028 CXX test/cpp_headers/ublk.o 00:05:18.028 CXX test/cpp_headers/util.o 00:05:18.028 CXX test/cpp_headers/uuid.o 00:05:18.028 CXX test/cpp_headers/version.o 00:05:18.028 CXX test/cpp_headers/vfio_user_pci.o 00:05:18.028 CXX test/cpp_headers/vfio_user_spec.o 00:05:18.028 CXX test/cpp_headers/vhost.o 00:05:18.028 CXX test/cpp_headers/vmd.o 00:05:18.028 CXX test/cpp_headers/xor.o 00:05:18.028 CXX test/cpp_headers/zipf.o 00:05:18.287 LINK cuse 00:05:20.212 LINK esnap 00:05:20.471 00:05:20.731 real 1m20.192s 00:05:20.731 user 6m10.716s 00:05:20.731 sys 1m10.849s 00:05:20.731 10:54:25 make -- common/autotest_common.sh@1128 -- $ xtrace_disable 00:05:20.731 10:54:25 make -- common/autotest_common.sh@10 -- $ set +x 00:05:20.731 ************************************ 00:05:20.731 END TEST make 00:05:20.731 ************************************ 00:05:20.731 10:54:26 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:20.731 10:54:26 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:20.731 10:54:26 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:20.731 10:54:26 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:20.731 10:54:26 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:20.731 10:54:26 -- pm/common@44 -- $ pid=6196 00:05:20.731 10:54:26 -- pm/common@50 -- $ kill -TERM 6196 00:05:20.731 10:54:26 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:20.731 10:54:26 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:20.731 10:54:26 -- pm/common@44 -- $ pid=6198 00:05:20.731 10:54:26 -- pm/common@50 -- $ kill -TERM 6198 00:05:20.731 10:54:26 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:05:20.731 10:54:26 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:05:20.731 10:54:26 -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:05:20.731 10:54:26 -- common/autotest_common.sh@1691 -- # lcov --version 00:05:20.731 10:54:26 -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:05:20.992 10:54:26 -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:05:20.992 10:54:26 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:20.992 10:54:26 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:20.992 10:54:26 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:20.992 10:54:26 -- scripts/common.sh@336 -- # IFS=.-: 00:05:20.992 10:54:26 -- scripts/common.sh@336 -- # read -ra ver1 00:05:20.992 10:54:26 -- scripts/common.sh@337 -- # IFS=.-: 00:05:20.992 10:54:26 -- scripts/common.sh@337 -- # read -ra ver2 00:05:20.992 10:54:26 -- scripts/common.sh@338 -- # local 'op=<' 00:05:20.992 10:54:26 -- scripts/common.sh@340 -- # ver1_l=2 00:05:20.992 10:54:26 -- scripts/common.sh@341 -- # ver2_l=1 00:05:20.992 10:54:26 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:20.992 10:54:26 -- scripts/common.sh@344 -- # case "$op" in 00:05:20.992 10:54:26 -- scripts/common.sh@345 -- # : 1 00:05:20.992 10:54:26 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:20.992 10:54:26 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:20.992 10:54:26 -- scripts/common.sh@365 -- # decimal 1 00:05:20.992 10:54:26 -- scripts/common.sh@353 -- # local d=1 00:05:20.992 10:54:26 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:20.992 10:54:26 -- scripts/common.sh@355 -- # echo 1 00:05:20.992 10:54:26 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:20.992 10:54:26 -- scripts/common.sh@366 -- # decimal 2 00:05:20.992 10:54:26 -- scripts/common.sh@353 -- # local d=2 00:05:20.992 10:54:26 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:20.992 10:54:26 -- scripts/common.sh@355 -- # echo 2 00:05:20.992 10:54:26 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:20.992 10:54:26 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:20.992 10:54:26 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:20.992 10:54:26 -- scripts/common.sh@368 -- # return 0 00:05:20.992 10:54:26 -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:20.992 10:54:26 -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:05:20.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:20.992 --rc genhtml_branch_coverage=1 00:05:20.992 --rc genhtml_function_coverage=1 00:05:20.992 --rc genhtml_legend=1 00:05:20.992 --rc geninfo_all_blocks=1 00:05:20.992 --rc geninfo_unexecuted_blocks=1 00:05:20.992 00:05:20.992 ' 00:05:20.992 10:54:26 -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:05:20.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:20.992 --rc genhtml_branch_coverage=1 00:05:20.992 --rc genhtml_function_coverage=1 00:05:20.992 --rc genhtml_legend=1 00:05:20.992 --rc geninfo_all_blocks=1 00:05:20.992 --rc geninfo_unexecuted_blocks=1 00:05:20.992 00:05:20.992 ' 00:05:20.992 10:54:26 -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:05:20.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:20.992 --rc genhtml_branch_coverage=1 00:05:20.992 --rc genhtml_function_coverage=1 00:05:20.992 --rc genhtml_legend=1 00:05:20.992 --rc geninfo_all_blocks=1 00:05:20.992 --rc geninfo_unexecuted_blocks=1 00:05:20.992 00:05:20.992 ' 00:05:20.992 10:54:26 -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:05:20.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:20.992 --rc genhtml_branch_coverage=1 00:05:20.992 --rc genhtml_function_coverage=1 00:05:20.992 --rc genhtml_legend=1 00:05:20.992 --rc geninfo_all_blocks=1 00:05:20.992 --rc geninfo_unexecuted_blocks=1 00:05:20.992 00:05:20.992 ' 00:05:20.992 10:54:26 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:20.992 10:54:26 -- nvmf/common.sh@7 -- # uname -s 00:05:20.992 10:54:26 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:20.992 10:54:26 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:20.992 10:54:26 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:20.992 10:54:26 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:20.992 10:54:26 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:20.992 10:54:26 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:20.992 10:54:26 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:20.992 10:54:26 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:20.993 10:54:26 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:20.993 10:54:26 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:20.993 10:54:26 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:05:20.993 10:54:26 -- nvmf/common.sh@18 -- # NVME_HOSTID=74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:05:20.993 10:54:26 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:20.993 10:54:26 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:20.993 10:54:26 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:20.993 10:54:26 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:20.993 10:54:26 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:20.993 10:54:26 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:20.993 10:54:26 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:20.993 10:54:26 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:20.993 10:54:26 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:20.993 10:54:26 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:20.993 10:54:26 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:20.993 10:54:26 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:20.993 10:54:26 -- paths/export.sh@5 -- # export PATH 00:05:20.993 10:54:26 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:20.993 10:54:26 -- nvmf/common.sh@51 -- # : 0 00:05:20.993 10:54:26 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:20.993 10:54:26 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:20.993 10:54:26 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:20.993 10:54:26 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:20.993 10:54:26 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:20.993 10:54:26 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:20.993 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:20.993 10:54:26 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:20.993 10:54:26 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:20.993 10:54:26 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:20.993 10:54:26 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:20.993 10:54:26 -- spdk/autotest.sh@32 -- # uname -s 00:05:20.993 10:54:26 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:20.993 10:54:26 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:20.993 10:54:26 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:20.993 10:54:26 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:20.993 10:54:26 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:20.993 10:54:26 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:20.993 10:54:26 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:20.993 10:54:26 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:20.993 10:54:26 -- spdk/autotest.sh@48 -- # udevadm_pid=66919 00:05:20.993 10:54:26 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:20.993 10:54:26 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:20.993 10:54:26 -- pm/common@17 -- # local monitor 00:05:20.993 10:54:26 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:20.993 10:54:26 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:20.993 10:54:26 -- pm/common@25 -- # sleep 1 00:05:20.993 10:54:26 -- pm/common@21 -- # date +%s 00:05:20.993 10:54:26 -- pm/common@21 -- # date +%s 00:05:20.993 10:54:26 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1730199266 00:05:20.993 10:54:26 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1730199266 00:05:20.993 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1730199266_collect-cpu-load.pm.log 00:05:20.993 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1730199266_collect-vmstat.pm.log 00:05:21.933 10:54:27 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:21.933 10:54:27 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:21.933 10:54:27 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:21.933 10:54:27 -- common/autotest_common.sh@10 -- # set +x 00:05:21.933 10:54:27 -- spdk/autotest.sh@59 -- # create_test_list 00:05:21.933 10:54:27 -- common/autotest_common.sh@750 -- # xtrace_disable 00:05:21.933 10:54:27 -- common/autotest_common.sh@10 -- # set +x 00:05:22.193 10:54:27 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:22.193 10:54:27 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:22.193 10:54:27 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:22.193 10:54:27 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:22.193 10:54:27 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:22.193 10:54:27 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:22.193 10:54:27 -- common/autotest_common.sh@1455 -- # uname 00:05:22.193 10:54:27 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:05:22.193 10:54:27 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:22.193 10:54:27 -- common/autotest_common.sh@1475 -- # uname 00:05:22.193 10:54:27 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:05:22.193 10:54:27 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:22.193 10:54:27 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:22.193 lcov: LCOV version 1.15 00:05:22.193 10:54:27 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:05:37.090 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:37.090 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:52.019 10:54:54 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:52.019 10:54:54 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:52.019 10:54:54 -- common/autotest_common.sh@10 -- # set +x 00:05:52.019 10:54:54 -- spdk/autotest.sh@78 -- # rm -f 00:05:52.019 10:54:54 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:52.019 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:52.019 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:52.019 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:52.019 10:54:55 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:52.019 10:54:55 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:05:52.019 10:54:55 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:05:52.019 10:54:55 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:05:52.019 10:54:55 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:52.019 10:54:55 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:05:52.019 10:54:55 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:05:52.019 10:54:55 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:52.019 10:54:55 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n1 00:05:52.019 10:54:55 -- common/autotest_common.sh@1648 -- # local device=nvme1n1 00:05:52.019 10:54:55 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:52.019 10:54:55 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n2 00:05:52.019 10:54:55 -- common/autotest_common.sh@1648 -- # local device=nvme1n2 00:05:52.019 10:54:55 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:52.019 10:54:55 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n3 00:05:52.019 10:54:55 -- common/autotest_common.sh@1648 -- # local device=nvme1n3 00:05:52.019 10:54:55 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:05:52.019 10:54:55 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:52.019 10:54:55 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:52.019 10:54:55 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:52.019 10:54:55 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:52.019 10:54:55 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:52.019 10:54:55 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:52.019 10:54:55 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:52.019 No valid GPT data, bailing 00:05:52.019 10:54:55 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:52.019 10:54:55 -- scripts/common.sh@394 -- # pt= 00:05:52.019 10:54:55 -- scripts/common.sh@395 -- # return 1 00:05:52.019 10:54:55 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:52.019 1+0 records in 00:05:52.019 1+0 records out 00:05:52.019 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00668402 s, 157 MB/s 00:05:52.019 10:54:55 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:52.019 10:54:55 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:52.019 10:54:55 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:52.019 10:54:55 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:52.019 10:54:55 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:52.019 No valid GPT data, bailing 00:05:52.019 10:54:55 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:52.019 10:54:55 -- scripts/common.sh@394 -- # pt= 00:05:52.019 10:54:55 -- scripts/common.sh@395 -- # return 1 00:05:52.019 10:54:55 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:52.019 1+0 records in 00:05:52.019 1+0 records out 00:05:52.019 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00644982 s, 163 MB/s 00:05:52.019 10:54:56 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:52.019 10:54:56 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:52.019 10:54:56 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:05:52.019 10:54:56 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:05:52.019 10:54:56 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:05:52.019 No valid GPT data, bailing 00:05:52.019 10:54:56 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:05:52.019 10:54:56 -- scripts/common.sh@394 -- # pt= 00:05:52.019 10:54:56 -- scripts/common.sh@395 -- # return 1 00:05:52.019 10:54:56 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:05:52.019 1+0 records in 00:05:52.019 1+0 records out 00:05:52.019 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00660523 s, 159 MB/s 00:05:52.019 10:54:56 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:52.019 10:54:56 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:52.019 10:54:56 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:05:52.019 10:54:56 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:05:52.019 10:54:56 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:05:52.019 No valid GPT data, bailing 00:05:52.019 10:54:56 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:05:52.019 10:54:56 -- scripts/common.sh@394 -- # pt= 00:05:52.019 10:54:56 -- scripts/common.sh@395 -- # return 1 00:05:52.019 10:54:56 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:05:52.019 1+0 records in 00:05:52.019 1+0 records out 00:05:52.019 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00636365 s, 165 MB/s 00:05:52.020 10:54:56 -- spdk/autotest.sh@105 -- # sync 00:05:52.020 10:54:56 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:52.020 10:54:56 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:52.020 10:54:56 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:54.556 10:54:59 -- spdk/autotest.sh@111 -- # uname -s 00:05:54.556 10:54:59 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:54.556 10:54:59 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:54.556 10:54:59 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:54.814 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:55.074 Hugepages 00:05:55.074 node hugesize free / total 00:05:55.074 node0 1048576kB 0 / 0 00:05:55.074 node0 2048kB 0 / 0 00:05:55.074 00:05:55.074 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:55.074 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:55.074 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:05:55.333 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:05:55.333 10:55:00 -- spdk/autotest.sh@117 -- # uname -s 00:05:55.333 10:55:00 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:55.333 10:55:00 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:55.333 10:55:00 -- common/autotest_common.sh@1514 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:56.272 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:56.272 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:56.272 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:56.272 10:55:01 -- common/autotest_common.sh@1515 -- # sleep 1 00:05:57.211 10:55:02 -- common/autotest_common.sh@1516 -- # bdfs=() 00:05:57.211 10:55:02 -- common/autotest_common.sh@1516 -- # local bdfs 00:05:57.211 10:55:02 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:05:57.211 10:55:02 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:05:57.211 10:55:02 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:57.211 10:55:02 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:57.211 10:55:02 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:57.211 10:55:02 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:57.211 10:55:02 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:57.471 10:55:02 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:57.471 10:55:02 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:57.471 10:55:02 -- common/autotest_common.sh@1520 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:58.042 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:58.042 Waiting for block devices as requested 00:05:58.042 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:58.042 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:58.302 10:55:03 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:58.302 10:55:03 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # grep 0000:00:10.0/nvme/nvme 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme1 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:58.302 10:55:03 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:58.302 10:55:03 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:58.302 10:55:03 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1541 -- # continue 00:05:58.302 10:55:03 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:58.302 10:55:03 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # grep 0000:00:11.0/nvme/nvme 00:05:58.302 10:55:03 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:58.302 10:55:03 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:58.302 10:55:03 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:58.302 10:55:03 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:58.302 10:55:03 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:58.302 10:55:03 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:58.302 10:55:03 -- common/autotest_common.sh@1541 -- # continue 00:05:58.302 10:55:03 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:58.302 10:55:03 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:58.302 10:55:03 -- common/autotest_common.sh@10 -- # set +x 00:05:58.302 10:55:03 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:58.302 10:55:03 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:58.302 10:55:03 -- common/autotest_common.sh@10 -- # set +x 00:05:58.302 10:55:03 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:59.242 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:59.242 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:59.242 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:59.242 10:55:04 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:59.242 10:55:04 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:59.242 10:55:04 -- common/autotest_common.sh@10 -- # set +x 00:05:59.516 10:55:04 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:59.516 10:55:04 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:05:59.516 10:55:04 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:05:59.516 10:55:04 -- common/autotest_common.sh@1561 -- # bdfs=() 00:05:59.516 10:55:04 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:05:59.516 10:55:04 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:05:59.516 10:55:04 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:05:59.516 10:55:04 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:05:59.516 10:55:04 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:59.516 10:55:04 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:59.516 10:55:04 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:59.516 10:55:04 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:59.516 10:55:04 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:59.516 10:55:04 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:59.516 10:55:04 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:59.516 10:55:04 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:59.516 10:55:04 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:59.516 10:55:04 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:59.516 10:55:04 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:59.516 10:55:04 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:59.516 10:55:04 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:59.516 10:55:04 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:59.516 10:55:04 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:59.516 10:55:04 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:05:59.516 10:55:04 -- common/autotest_common.sh@1570 -- # return 0 00:05:59.516 10:55:04 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:05:59.516 10:55:04 -- common/autotest_common.sh@1578 -- # return 0 00:05:59.516 10:55:04 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:59.516 10:55:04 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:59.516 10:55:04 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:59.516 10:55:04 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:59.516 10:55:04 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:59.516 10:55:04 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:59.516 10:55:04 -- common/autotest_common.sh@10 -- # set +x 00:05:59.516 10:55:04 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:59.516 10:55:04 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:59.516 10:55:04 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:05:59.516 10:55:04 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:05:59.516 10:55:04 -- common/autotest_common.sh@10 -- # set +x 00:05:59.516 ************************************ 00:05:59.516 START TEST env 00:05:59.516 ************************************ 00:05:59.516 10:55:04 env -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:59.790 * Looking for test storage... 00:05:59.790 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1691 -- # lcov --version 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:05:59.790 10:55:05 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:59.790 10:55:05 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:59.790 10:55:05 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:59.790 10:55:05 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:59.790 10:55:05 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:59.790 10:55:05 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:59.790 10:55:05 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:59.790 10:55:05 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:59.790 10:55:05 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:59.790 10:55:05 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:59.790 10:55:05 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:59.790 10:55:05 env -- scripts/common.sh@344 -- # case "$op" in 00:05:59.790 10:55:05 env -- scripts/common.sh@345 -- # : 1 00:05:59.790 10:55:05 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:59.790 10:55:05 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:59.790 10:55:05 env -- scripts/common.sh@365 -- # decimal 1 00:05:59.790 10:55:05 env -- scripts/common.sh@353 -- # local d=1 00:05:59.790 10:55:05 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:59.790 10:55:05 env -- scripts/common.sh@355 -- # echo 1 00:05:59.790 10:55:05 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:59.790 10:55:05 env -- scripts/common.sh@366 -- # decimal 2 00:05:59.790 10:55:05 env -- scripts/common.sh@353 -- # local d=2 00:05:59.790 10:55:05 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:59.790 10:55:05 env -- scripts/common.sh@355 -- # echo 2 00:05:59.790 10:55:05 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:59.790 10:55:05 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:59.790 10:55:05 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:59.790 10:55:05 env -- scripts/common.sh@368 -- # return 0 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:05:59.790 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.790 --rc genhtml_branch_coverage=1 00:05:59.790 --rc genhtml_function_coverage=1 00:05:59.790 --rc genhtml_legend=1 00:05:59.790 --rc geninfo_all_blocks=1 00:05:59.790 --rc geninfo_unexecuted_blocks=1 00:05:59.790 00:05:59.790 ' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:05:59.790 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.790 --rc genhtml_branch_coverage=1 00:05:59.790 --rc genhtml_function_coverage=1 00:05:59.790 --rc genhtml_legend=1 00:05:59.790 --rc geninfo_all_blocks=1 00:05:59.790 --rc geninfo_unexecuted_blocks=1 00:05:59.790 00:05:59.790 ' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:05:59.790 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.790 --rc genhtml_branch_coverage=1 00:05:59.790 --rc genhtml_function_coverage=1 00:05:59.790 --rc genhtml_legend=1 00:05:59.790 --rc geninfo_all_blocks=1 00:05:59.790 --rc geninfo_unexecuted_blocks=1 00:05:59.790 00:05:59.790 ' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:05:59.790 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:59.790 --rc genhtml_branch_coverage=1 00:05:59.790 --rc genhtml_function_coverage=1 00:05:59.790 --rc genhtml_legend=1 00:05:59.790 --rc geninfo_all_blocks=1 00:05:59.790 --rc geninfo_unexecuted_blocks=1 00:05:59.790 00:05:59.790 ' 00:05:59.790 10:55:05 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:05:59.790 10:55:05 env -- common/autotest_common.sh@1109 -- # xtrace_disable 00:05:59.790 10:55:05 env -- common/autotest_common.sh@10 -- # set +x 00:05:59.790 ************************************ 00:05:59.790 START TEST env_memory 00:05:59.790 ************************************ 00:05:59.790 10:55:05 env.env_memory -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:59.790 00:05:59.790 00:05:59.790 CUnit - A unit testing framework for C - Version 2.1-3 00:05:59.790 http://cunit.sourceforge.net/ 00:05:59.790 00:05:59.790 00:05:59.790 Suite: memory 00:05:59.790 Test: alloc and free memory map ...[2024-10-29 10:55:05.199741] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:59.790 passed 00:05:59.790 Test: mem map translation ...[2024-10-29 10:55:05.244167] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:59.790 [2024-10-29 10:55:05.244214] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:59.790 [2024-10-29 10:55:05.244274] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:59.790 [2024-10-29 10:55:05.244293] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:06:00.051 passed 00:06:00.051 Test: mem map registration ...[2024-10-29 10:55:05.310355] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:06:00.051 [2024-10-29 10:55:05.310402] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:06:00.051 passed 00:06:00.051 Test: mem map adjacent registrations ...passed 00:06:00.051 00:06:00.051 Run Summary: Type Total Ran Passed Failed Inactive 00:06:00.051 suites 1 1 n/a 0 0 00:06:00.051 tests 4 4 4 0 0 00:06:00.051 asserts 152 152 152 0 n/a 00:06:00.051 00:06:00.051 Elapsed time = 0.237 seconds 00:06:00.051 00:06:00.051 real 0m0.282s 00:06:00.051 user 0m0.249s 00:06:00.051 sys 0m0.024s 00:06:00.051 10:55:05 env.env_memory -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:00.051 10:55:05 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:06:00.051 ************************************ 00:06:00.051 END TEST env_memory 00:06:00.051 ************************************ 00:06:00.051 10:55:05 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:00.051 10:55:05 env -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:00.051 10:55:05 env -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:00.051 10:55:05 env -- common/autotest_common.sh@10 -- # set +x 00:06:00.051 ************************************ 00:06:00.051 START TEST env_vtophys 00:06:00.051 ************************************ 00:06:00.051 10:55:05 env.env_vtophys -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:06:00.051 EAL: lib.eal log level changed from notice to debug 00:06:00.051 EAL: Detected lcore 0 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 1 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 2 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 3 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 4 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 5 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 6 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 7 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 8 as core 0 on socket 0 00:06:00.051 EAL: Detected lcore 9 as core 0 on socket 0 00:06:00.051 EAL: Maximum logical cores by configuration: 128 00:06:00.051 EAL: Detected CPU lcores: 10 00:06:00.051 EAL: Detected NUMA nodes: 1 00:06:00.051 EAL: Checking presence of .so 'librte_eal.so.24.0' 00:06:00.051 EAL: Detected shared linkage of DPDK 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24.0 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24.0 00:06:00.051 EAL: Registered [vdev] bus. 00:06:00.051 EAL: bus.vdev log level changed from disabled to notice 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24.0 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24.0 00:06:00.051 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:06:00.051 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:06:00.051 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:06:00.051 EAL: No shared files mode enabled, IPC will be disabled 00:06:00.311 EAL: No shared files mode enabled, IPC is disabled 00:06:00.311 EAL: Selected IOVA mode 'PA' 00:06:00.311 EAL: Probing VFIO support... 00:06:00.311 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:00.311 EAL: VFIO modules not loaded, skipping VFIO support... 00:06:00.311 EAL: Ask a virtual area of 0x2e000 bytes 00:06:00.311 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:06:00.311 EAL: Setting up physically contiguous memory... 00:06:00.311 EAL: Setting maximum number of open files to 524288 00:06:00.311 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:06:00.311 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:06:00.311 EAL: Ask a virtual area of 0x61000 bytes 00:06:00.311 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:06:00.311 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:00.311 EAL: Ask a virtual area of 0x400000000 bytes 00:06:00.311 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:06:00.311 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:06:00.311 EAL: Ask a virtual area of 0x61000 bytes 00:06:00.311 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:06:00.311 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:00.311 EAL: Ask a virtual area of 0x400000000 bytes 00:06:00.311 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:06:00.311 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:06:00.311 EAL: Ask a virtual area of 0x61000 bytes 00:06:00.311 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:06:00.312 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:00.312 EAL: Ask a virtual area of 0x400000000 bytes 00:06:00.312 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:06:00.312 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:06:00.312 EAL: Ask a virtual area of 0x61000 bytes 00:06:00.312 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:06:00.312 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:06:00.312 EAL: Ask a virtual area of 0x400000000 bytes 00:06:00.312 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:06:00.312 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:06:00.312 EAL: Hugepages will be freed exactly as allocated. 00:06:00.312 EAL: No shared files mode enabled, IPC is disabled 00:06:00.312 EAL: No shared files mode enabled, IPC is disabled 00:06:00.312 EAL: TSC frequency is ~2290000 KHz 00:06:00.312 EAL: Main lcore 0 is ready (tid=7fa8907b6a40;cpuset=[0]) 00:06:00.312 EAL: Trying to obtain current memory policy. 00:06:00.312 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.312 EAL: Restoring previous memory policy: 0 00:06:00.312 EAL: request: mp_malloc_sync 00:06:00.312 EAL: No shared files mode enabled, IPC is disabled 00:06:00.312 EAL: Heap on socket 0 was expanded by 2MB 00:06:00.312 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:06:00.312 EAL: No shared files mode enabled, IPC is disabled 00:06:00.312 EAL: No PCI address specified using 'addr=' in: bus=pci 00:06:00.312 EAL: Mem event callback 'spdk:(nil)' registered 00:06:00.312 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:06:00.312 00:06:00.312 00:06:00.312 CUnit - A unit testing framework for C - Version 2.1-3 00:06:00.312 http://cunit.sourceforge.net/ 00:06:00.312 00:06:00.312 00:06:00.312 Suite: components_suite 00:06:00.572 Test: vtophys_malloc_test ...passed 00:06:00.572 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:06:00.572 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.572 EAL: Restoring previous memory policy: 4 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was expanded by 4MB 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was shrunk by 4MB 00:06:00.572 EAL: Trying to obtain current memory policy. 00:06:00.572 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.572 EAL: Restoring previous memory policy: 4 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was expanded by 6MB 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was shrunk by 6MB 00:06:00.572 EAL: Trying to obtain current memory policy. 00:06:00.572 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.572 EAL: Restoring previous memory policy: 4 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was expanded by 10MB 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was shrunk by 10MB 00:06:00.572 EAL: Trying to obtain current memory policy. 00:06:00.572 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.572 EAL: Restoring previous memory policy: 4 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was expanded by 18MB 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was shrunk by 18MB 00:06:00.572 EAL: Trying to obtain current memory policy. 00:06:00.572 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.572 EAL: Restoring previous memory policy: 4 00:06:00.572 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.572 EAL: request: mp_malloc_sync 00:06:00.572 EAL: No shared files mode enabled, IPC is disabled 00:06:00.572 EAL: Heap on socket 0 was expanded by 34MB 00:06:00.573 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.573 EAL: request: mp_malloc_sync 00:06:00.573 EAL: No shared files mode enabled, IPC is disabled 00:06:00.573 EAL: Heap on socket 0 was shrunk by 34MB 00:06:00.573 EAL: Trying to obtain current memory policy. 00:06:00.573 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.573 EAL: Restoring previous memory policy: 4 00:06:00.573 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.573 EAL: request: mp_malloc_sync 00:06:00.573 EAL: No shared files mode enabled, IPC is disabled 00:06:00.573 EAL: Heap on socket 0 was expanded by 66MB 00:06:00.833 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.833 EAL: request: mp_malloc_sync 00:06:00.833 EAL: No shared files mode enabled, IPC is disabled 00:06:00.833 EAL: Heap on socket 0 was shrunk by 66MB 00:06:00.833 EAL: Trying to obtain current memory policy. 00:06:00.833 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.833 EAL: Restoring previous memory policy: 4 00:06:00.833 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.833 EAL: request: mp_malloc_sync 00:06:00.833 EAL: No shared files mode enabled, IPC is disabled 00:06:00.833 EAL: Heap on socket 0 was expanded by 130MB 00:06:00.833 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.833 EAL: request: mp_malloc_sync 00:06:00.833 EAL: No shared files mode enabled, IPC is disabled 00:06:00.833 EAL: Heap on socket 0 was shrunk by 130MB 00:06:00.833 EAL: Trying to obtain current memory policy. 00:06:00.833 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:00.833 EAL: Restoring previous memory policy: 4 00:06:00.833 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.833 EAL: request: mp_malloc_sync 00:06:00.833 EAL: No shared files mode enabled, IPC is disabled 00:06:00.833 EAL: Heap on socket 0 was expanded by 258MB 00:06:00.833 EAL: Calling mem event callback 'spdk:(nil)' 00:06:00.833 EAL: request: mp_malloc_sync 00:06:00.833 EAL: No shared files mode enabled, IPC is disabled 00:06:00.833 EAL: Heap on socket 0 was shrunk by 258MB 00:06:00.833 EAL: Trying to obtain current memory policy. 00:06:00.833 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:01.092 EAL: Restoring previous memory policy: 4 00:06:01.092 EAL: Calling mem event callback 'spdk:(nil)' 00:06:01.092 EAL: request: mp_malloc_sync 00:06:01.092 EAL: No shared files mode enabled, IPC is disabled 00:06:01.092 EAL: Heap on socket 0 was expanded by 514MB 00:06:01.092 EAL: Calling mem event callback 'spdk:(nil)' 00:06:01.092 EAL: request: mp_malloc_sync 00:06:01.092 EAL: No shared files mode enabled, IPC is disabled 00:06:01.092 EAL: Heap on socket 0 was shrunk by 514MB 00:06:01.092 EAL: Trying to obtain current memory policy. 00:06:01.092 EAL: Setting policy MPOL_PREFERRED for socket 0 00:06:01.351 EAL: Restoring previous memory policy: 4 00:06:01.351 EAL: Calling mem event callback 'spdk:(nil)' 00:06:01.351 EAL: request: mp_malloc_sync 00:06:01.351 EAL: No shared files mode enabled, IPC is disabled 00:06:01.351 EAL: Heap on socket 0 was expanded by 1026MB 00:06:01.612 EAL: Calling mem event callback 'spdk:(nil)' 00:06:01.612 passed 00:06:01.612 00:06:01.612 Run Summary: Type Total Ran Passed Failed Inactive 00:06:01.612 suites 1 1 n/a 0 0 00:06:01.612 tests 2 2 2 0 0 00:06:01.612 asserts 5092 5092 5092 0 n/a 00:06:01.612 00:06:01.612 Elapsed time = 1.346 seconds 00:06:01.612 EAL: request: mp_malloc_sync 00:06:01.612 EAL: No shared files mode enabled, IPC is disabled 00:06:01.612 EAL: Heap on socket 0 was shrunk by 1026MB 00:06:01.612 EAL: Calling mem event callback 'spdk:(nil)' 00:06:01.612 EAL: request: mp_malloc_sync 00:06:01.612 EAL: No shared files mode enabled, IPC is disabled 00:06:01.612 EAL: Heap on socket 0 was shrunk by 2MB 00:06:01.612 EAL: No shared files mode enabled, IPC is disabled 00:06:01.612 EAL: No shared files mode enabled, IPC is disabled 00:06:01.612 EAL: No shared files mode enabled, IPC is disabled 00:06:01.612 00:06:01.612 real 0m1.618s 00:06:01.612 user 0m0.755s 00:06:01.612 sys 0m0.730s 00:06:01.612 10:55:07 env.env_vtophys -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:01.612 10:55:07 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:06:01.612 ************************************ 00:06:01.612 END TEST env_vtophys 00:06:01.612 ************************************ 00:06:01.871 10:55:07 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:01.871 10:55:07 env -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:01.871 10:55:07 env -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:01.871 10:55:07 env -- common/autotest_common.sh@10 -- # set +x 00:06:01.871 ************************************ 00:06:01.871 START TEST env_pci 00:06:01.871 ************************************ 00:06:01.871 10:55:07 env.env_pci -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:06:01.871 00:06:01.871 00:06:01.871 CUnit - A unit testing framework for C - Version 2.1-3 00:06:01.871 http://cunit.sourceforge.net/ 00:06:01.871 00:06:01.871 00:06:01.871 Suite: pci 00:06:01.871 Test: pci_hook ...[2024-10-29 10:55:07.192682] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 69137 has claimed it 00:06:01.871 passed 00:06:01.871 00:06:01.871 EAL: Cannot find device (10000:00:01.0) 00:06:01.871 EAL: Failed to attach device on primary process 00:06:01.871 Run Summary: Type Total Ran Passed Failed Inactive 00:06:01.872 suites 1 1 n/a 0 0 00:06:01.872 tests 1 1 1 0 0 00:06:01.872 asserts 25 25 25 0 n/a 00:06:01.872 00:06:01.872 Elapsed time = 0.007 seconds 00:06:01.872 00:06:01.872 real 0m0.092s 00:06:01.872 user 0m0.045s 00:06:01.872 sys 0m0.045s 00:06:01.872 10:55:07 env.env_pci -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:01.872 10:55:07 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:06:01.872 ************************************ 00:06:01.872 END TEST env_pci 00:06:01.872 ************************************ 00:06:01.872 10:55:07 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:06:01.872 10:55:07 env -- env/env.sh@15 -- # uname 00:06:01.872 10:55:07 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:06:01.872 10:55:07 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:06:01.872 10:55:07 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:01.872 10:55:07 env -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:06:01.872 10:55:07 env -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:01.872 10:55:07 env -- common/autotest_common.sh@10 -- # set +x 00:06:01.872 ************************************ 00:06:01.872 START TEST env_dpdk_post_init 00:06:01.872 ************************************ 00:06:01.872 10:55:07 env.env_dpdk_post_init -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:06:02.135 EAL: Detected CPU lcores: 10 00:06:02.135 EAL: Detected NUMA nodes: 1 00:06:02.135 EAL: Detected shared linkage of DPDK 00:06:02.135 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:02.135 EAL: Selected IOVA mode 'PA' 00:06:02.135 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:02.135 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:06:02.135 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:06:02.135 Starting DPDK initialization... 00:06:02.135 Starting SPDK post initialization... 00:06:02.135 SPDK NVMe probe 00:06:02.135 Attaching to 0000:00:10.0 00:06:02.135 Attaching to 0000:00:11.0 00:06:02.135 Attached to 0000:00:10.0 00:06:02.135 Attached to 0000:00:11.0 00:06:02.135 Cleaning up... 00:06:02.135 00:06:02.135 real 0m0.261s 00:06:02.135 user 0m0.080s 00:06:02.135 sys 0m0.081s 00:06:02.135 10:55:07 env.env_dpdk_post_init -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:02.135 10:55:07 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:06:02.135 ************************************ 00:06:02.135 END TEST env_dpdk_post_init 00:06:02.135 ************************************ 00:06:02.395 10:55:07 env -- env/env.sh@26 -- # uname 00:06:02.395 10:55:07 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:06:02.395 10:55:07 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:02.395 10:55:07 env -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:02.395 10:55:07 env -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:02.395 10:55:07 env -- common/autotest_common.sh@10 -- # set +x 00:06:02.395 ************************************ 00:06:02.395 START TEST env_mem_callbacks 00:06:02.395 ************************************ 00:06:02.395 10:55:07 env.env_mem_callbacks -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:06:02.395 EAL: Detected CPU lcores: 10 00:06:02.395 EAL: Detected NUMA nodes: 1 00:06:02.395 EAL: Detected shared linkage of DPDK 00:06:02.395 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:06:02.395 EAL: Selected IOVA mode 'PA' 00:06:02.395 TELEMETRY: No legacy callbacks, legacy socket not created 00:06:02.395 00:06:02.395 00:06:02.395 CUnit - A unit testing framework for C - Version 2.1-3 00:06:02.395 http://cunit.sourceforge.net/ 00:06:02.395 00:06:02.395 00:06:02.395 Suite: memory 00:06:02.395 Test: test ... 00:06:02.395 register 0x200000200000 2097152 00:06:02.395 malloc 3145728 00:06:02.395 register 0x200000400000 4194304 00:06:02.395 buf 0x200000500000 len 3145728 PASSED 00:06:02.395 malloc 64 00:06:02.395 buf 0x2000004fff40 len 64 PASSED 00:06:02.395 malloc 4194304 00:06:02.395 register 0x200000800000 6291456 00:06:02.395 buf 0x200000a00000 len 4194304 PASSED 00:06:02.395 free 0x200000500000 3145728 00:06:02.395 free 0x2000004fff40 64 00:06:02.395 unregister 0x200000400000 4194304 PASSED 00:06:02.395 free 0x200000a00000 4194304 00:06:02.395 unregister 0x200000800000 6291456 PASSED 00:06:02.395 malloc 8388608 00:06:02.395 register 0x200000400000 10485760 00:06:02.395 buf 0x200000600000 len 8388608 PASSED 00:06:02.395 free 0x200000600000 8388608 00:06:02.395 unregister 0x200000400000 10485760 PASSED 00:06:02.395 passed 00:06:02.395 00:06:02.395 Run Summary: Type Total Ran Passed Failed Inactive 00:06:02.395 suites 1 1 n/a 0 0 00:06:02.395 tests 1 1 1 0 0 00:06:02.395 asserts 15 15 15 0 n/a 00:06:02.395 00:06:02.395 Elapsed time = 0.011 seconds 00:06:02.395 00:06:02.395 real 0m0.197s 00:06:02.395 user 0m0.035s 00:06:02.395 sys 0m0.060s 00:06:02.395 10:55:07 env.env_mem_callbacks -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:02.395 10:55:07 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:06:02.395 ************************************ 00:06:02.395 END TEST env_mem_callbacks 00:06:02.395 ************************************ 00:06:02.655 00:06:02.655 real 0m3.008s 00:06:02.655 user 0m1.397s 00:06:02.655 sys 0m1.288s 00:06:02.655 10:55:07 env -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:02.655 10:55:07 env -- common/autotest_common.sh@10 -- # set +x 00:06:02.655 ************************************ 00:06:02.655 END TEST env 00:06:02.655 ************************************ 00:06:02.655 10:55:07 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:02.655 10:55:07 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:02.655 10:55:07 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:02.655 10:55:07 -- common/autotest_common.sh@10 -- # set +x 00:06:02.655 ************************************ 00:06:02.655 START TEST rpc 00:06:02.655 ************************************ 00:06:02.655 10:55:07 rpc -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:06:02.655 * Looking for test storage... 00:06:02.655 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:02.655 10:55:08 rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:02.655 10:55:08 rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:06:02.655 10:55:08 rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:02.916 10:55:08 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:02.916 10:55:08 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:02.916 10:55:08 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:02.916 10:55:08 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:02.916 10:55:08 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:02.916 10:55:08 rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:02.916 10:55:08 rpc -- scripts/common.sh@345 -- # : 1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:02.916 10:55:08 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:02.916 10:55:08 rpc -- scripts/common.sh@365 -- # decimal 1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@353 -- # local d=1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:02.916 10:55:08 rpc -- scripts/common.sh@355 -- # echo 1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:02.916 10:55:08 rpc -- scripts/common.sh@366 -- # decimal 2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@353 -- # local d=2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:02.916 10:55:08 rpc -- scripts/common.sh@355 -- # echo 2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:02.916 10:55:08 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:02.916 10:55:08 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:02.916 10:55:08 rpc -- scripts/common.sh@368 -- # return 0 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:02.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:02.916 --rc genhtml_branch_coverage=1 00:06:02.916 --rc genhtml_function_coverage=1 00:06:02.916 --rc genhtml_legend=1 00:06:02.916 --rc geninfo_all_blocks=1 00:06:02.916 --rc geninfo_unexecuted_blocks=1 00:06:02.916 00:06:02.916 ' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:02.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:02.916 --rc genhtml_branch_coverage=1 00:06:02.916 --rc genhtml_function_coverage=1 00:06:02.916 --rc genhtml_legend=1 00:06:02.916 --rc geninfo_all_blocks=1 00:06:02.916 --rc geninfo_unexecuted_blocks=1 00:06:02.916 00:06:02.916 ' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:02.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:02.916 --rc genhtml_branch_coverage=1 00:06:02.916 --rc genhtml_function_coverage=1 00:06:02.916 --rc genhtml_legend=1 00:06:02.916 --rc geninfo_all_blocks=1 00:06:02.916 --rc geninfo_unexecuted_blocks=1 00:06:02.916 00:06:02.916 ' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:02.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:02.916 --rc genhtml_branch_coverage=1 00:06:02.916 --rc genhtml_function_coverage=1 00:06:02.916 --rc genhtml_legend=1 00:06:02.916 --rc geninfo_all_blocks=1 00:06:02.916 --rc geninfo_unexecuted_blocks=1 00:06:02.916 00:06:02.916 ' 00:06:02.916 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:02.916 10:55:08 rpc -- rpc/rpc.sh@65 -- # spdk_pid=69264 00:06:02.916 10:55:08 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:02.916 10:55:08 rpc -- rpc/rpc.sh@67 -- # waitforlisten 69264 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@833 -- # '[' -z 69264 ']' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:02.916 10:55:08 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:02.916 10:55:08 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:02.916 [2024-10-29 10:55:08.344599] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:02.916 [2024-10-29 10:55:08.344746] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69264 ] 00:06:03.177 [2024-10-29 10:55:08.517638] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:03.177 [2024-10-29 10:55:08.541578] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:06:03.177 [2024-10-29 10:55:08.541640] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 69264' to capture a snapshot of events at runtime. 00:06:03.177 [2024-10-29 10:55:08.541659] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:03.177 [2024-10-29 10:55:08.541688] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:03.177 [2024-10-29 10:55:08.541701] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid69264 for offline analysis/debug. 00:06:03.177 [2024-10-29 10:55:08.542063] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:03.746 10:55:09 rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:03.746 10:55:09 rpc -- common/autotest_common.sh@866 -- # return 0 00:06:03.746 10:55:09 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:03.746 10:55:09 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:06:03.746 10:55:09 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:06:03.746 10:55:09 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:06:03.746 10:55:09 rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:03.746 10:55:09 rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:03.746 10:55:09 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.746 ************************************ 00:06:03.746 START TEST rpc_integrity 00:06:03.746 ************************************ 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@1127 -- # rpc_integrity 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:03.746 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:03.746 { 00:06:03.746 "name": "Malloc0", 00:06:03.746 "aliases": [ 00:06:03.746 "bc676d77-d044-4d55-84f1-b2b86c9e96c8" 00:06:03.746 ], 00:06:03.746 "product_name": "Malloc disk", 00:06:03.746 "block_size": 512, 00:06:03.746 "num_blocks": 16384, 00:06:03.746 "uuid": "bc676d77-d044-4d55-84f1-b2b86c9e96c8", 00:06:03.746 "assigned_rate_limits": { 00:06:03.746 "rw_ios_per_sec": 0, 00:06:03.746 "rw_mbytes_per_sec": 0, 00:06:03.746 "r_mbytes_per_sec": 0, 00:06:03.746 "w_mbytes_per_sec": 0 00:06:03.746 }, 00:06:03.746 "claimed": false, 00:06:03.746 "zoned": false, 00:06:03.746 "supported_io_types": { 00:06:03.746 "read": true, 00:06:03.746 "write": true, 00:06:03.746 "unmap": true, 00:06:03.746 "flush": true, 00:06:03.746 "reset": true, 00:06:03.746 "nvme_admin": false, 00:06:03.746 "nvme_io": false, 00:06:03.746 "nvme_io_md": false, 00:06:03.746 "write_zeroes": true, 00:06:03.746 "zcopy": true, 00:06:03.746 "get_zone_info": false, 00:06:03.746 "zone_management": false, 00:06:03.746 "zone_append": false, 00:06:03.746 "compare": false, 00:06:03.746 "compare_and_write": false, 00:06:03.746 "abort": true, 00:06:03.746 "seek_hole": false, 00:06:03.746 "seek_data": false, 00:06:03.746 "copy": true, 00:06:03.746 "nvme_iov_md": false 00:06:03.746 }, 00:06:03.746 "memory_domains": [ 00:06:03.746 { 00:06:03.746 "dma_device_id": "system", 00:06:03.746 "dma_device_type": 1 00:06:03.746 }, 00:06:03.746 { 00:06:03.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:03.746 "dma_device_type": 2 00:06:03.746 } 00:06:03.746 ], 00:06:03.746 "driver_specific": {} 00:06:03.746 } 00:06:03.746 ]' 00:06:03.746 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:04.006 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:04.006 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.006 [2024-10-29 10:55:09.283060] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:06:04.006 [2024-10-29 10:55:09.283121] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:04.006 [2024-10-29 10:55:09.283153] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:06:04.006 [2024-10-29 10:55:09.283164] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:04.006 [2024-10-29 10:55:09.285486] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:04.006 [2024-10-29 10:55:09.285522] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:04.006 Passthru0 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.006 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.006 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.006 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:04.006 { 00:06:04.006 "name": "Malloc0", 00:06:04.006 "aliases": [ 00:06:04.006 "bc676d77-d044-4d55-84f1-b2b86c9e96c8" 00:06:04.006 ], 00:06:04.006 "product_name": "Malloc disk", 00:06:04.006 "block_size": 512, 00:06:04.006 "num_blocks": 16384, 00:06:04.006 "uuid": "bc676d77-d044-4d55-84f1-b2b86c9e96c8", 00:06:04.006 "assigned_rate_limits": { 00:06:04.006 "rw_ios_per_sec": 0, 00:06:04.006 "rw_mbytes_per_sec": 0, 00:06:04.006 "r_mbytes_per_sec": 0, 00:06:04.006 "w_mbytes_per_sec": 0 00:06:04.006 }, 00:06:04.006 "claimed": true, 00:06:04.006 "claim_type": "exclusive_write", 00:06:04.006 "zoned": false, 00:06:04.006 "supported_io_types": { 00:06:04.006 "read": true, 00:06:04.006 "write": true, 00:06:04.007 "unmap": true, 00:06:04.007 "flush": true, 00:06:04.007 "reset": true, 00:06:04.007 "nvme_admin": false, 00:06:04.007 "nvme_io": false, 00:06:04.007 "nvme_io_md": false, 00:06:04.007 "write_zeroes": true, 00:06:04.007 "zcopy": true, 00:06:04.007 "get_zone_info": false, 00:06:04.007 "zone_management": false, 00:06:04.007 "zone_append": false, 00:06:04.007 "compare": false, 00:06:04.007 "compare_and_write": false, 00:06:04.007 "abort": true, 00:06:04.007 "seek_hole": false, 00:06:04.007 "seek_data": false, 00:06:04.007 "copy": true, 00:06:04.007 "nvme_iov_md": false 00:06:04.007 }, 00:06:04.007 "memory_domains": [ 00:06:04.007 { 00:06:04.007 "dma_device_id": "system", 00:06:04.007 "dma_device_type": 1 00:06:04.007 }, 00:06:04.007 { 00:06:04.007 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.007 "dma_device_type": 2 00:06:04.007 } 00:06:04.007 ], 00:06:04.007 "driver_specific": {} 00:06:04.007 }, 00:06:04.007 { 00:06:04.007 "name": "Passthru0", 00:06:04.007 "aliases": [ 00:06:04.007 "c6d3a7a3-ffdb-5cbb-8229-c50783134178" 00:06:04.007 ], 00:06:04.007 "product_name": "passthru", 00:06:04.007 "block_size": 512, 00:06:04.007 "num_blocks": 16384, 00:06:04.007 "uuid": "c6d3a7a3-ffdb-5cbb-8229-c50783134178", 00:06:04.007 "assigned_rate_limits": { 00:06:04.007 "rw_ios_per_sec": 0, 00:06:04.007 "rw_mbytes_per_sec": 0, 00:06:04.007 "r_mbytes_per_sec": 0, 00:06:04.007 "w_mbytes_per_sec": 0 00:06:04.007 }, 00:06:04.007 "claimed": false, 00:06:04.007 "zoned": false, 00:06:04.007 "supported_io_types": { 00:06:04.007 "read": true, 00:06:04.007 "write": true, 00:06:04.007 "unmap": true, 00:06:04.007 "flush": true, 00:06:04.007 "reset": true, 00:06:04.007 "nvme_admin": false, 00:06:04.007 "nvme_io": false, 00:06:04.007 "nvme_io_md": false, 00:06:04.007 "write_zeroes": true, 00:06:04.007 "zcopy": true, 00:06:04.007 "get_zone_info": false, 00:06:04.007 "zone_management": false, 00:06:04.007 "zone_append": false, 00:06:04.007 "compare": false, 00:06:04.007 "compare_and_write": false, 00:06:04.007 "abort": true, 00:06:04.007 "seek_hole": false, 00:06:04.007 "seek_data": false, 00:06:04.007 "copy": true, 00:06:04.007 "nvme_iov_md": false 00:06:04.007 }, 00:06:04.007 "memory_domains": [ 00:06:04.007 { 00:06:04.007 "dma_device_id": "system", 00:06:04.007 "dma_device_type": 1 00:06:04.007 }, 00:06:04.007 { 00:06:04.007 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.007 "dma_device_type": 2 00:06:04.007 } 00:06:04.007 ], 00:06:04.007 "driver_specific": { 00:06:04.007 "passthru": { 00:06:04.007 "name": "Passthru0", 00:06:04.007 "base_bdev_name": "Malloc0" 00:06:04.007 } 00:06:04.007 } 00:06:04.007 } 00:06:04.007 ]' 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:04.007 ************************************ 00:06:04.007 END TEST rpc_integrity 00:06:04.007 ************************************ 00:06:04.007 10:55:09 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:04.007 00:06:04.007 real 0m0.336s 00:06:04.007 user 0m0.196s 00:06:04.007 sys 0m0.057s 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:04.007 10:55:09 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.007 10:55:09 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:06:04.007 10:55:09 rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:04.007 10:55:09 rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:04.007 10:55:09 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:04.266 ************************************ 00:06:04.266 START TEST rpc_plugins 00:06:04.266 ************************************ 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@1127 -- # rpc_plugins 00:06:04.266 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.266 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:06:04.266 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:04.266 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.266 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:06:04.266 { 00:06:04.266 "name": "Malloc1", 00:06:04.266 "aliases": [ 00:06:04.266 "8a4e0378-4089-4df9-94a6-d2fe7a070195" 00:06:04.266 ], 00:06:04.266 "product_name": "Malloc disk", 00:06:04.266 "block_size": 4096, 00:06:04.266 "num_blocks": 256, 00:06:04.266 "uuid": "8a4e0378-4089-4df9-94a6-d2fe7a070195", 00:06:04.266 "assigned_rate_limits": { 00:06:04.266 "rw_ios_per_sec": 0, 00:06:04.266 "rw_mbytes_per_sec": 0, 00:06:04.266 "r_mbytes_per_sec": 0, 00:06:04.266 "w_mbytes_per_sec": 0 00:06:04.266 }, 00:06:04.266 "claimed": false, 00:06:04.266 "zoned": false, 00:06:04.266 "supported_io_types": { 00:06:04.266 "read": true, 00:06:04.266 "write": true, 00:06:04.266 "unmap": true, 00:06:04.266 "flush": true, 00:06:04.266 "reset": true, 00:06:04.266 "nvme_admin": false, 00:06:04.266 "nvme_io": false, 00:06:04.266 "nvme_io_md": false, 00:06:04.266 "write_zeroes": true, 00:06:04.266 "zcopy": true, 00:06:04.267 "get_zone_info": false, 00:06:04.267 "zone_management": false, 00:06:04.267 "zone_append": false, 00:06:04.267 "compare": false, 00:06:04.267 "compare_and_write": false, 00:06:04.267 "abort": true, 00:06:04.267 "seek_hole": false, 00:06:04.267 "seek_data": false, 00:06:04.267 "copy": true, 00:06:04.267 "nvme_iov_md": false 00:06:04.267 }, 00:06:04.267 "memory_domains": [ 00:06:04.267 { 00:06:04.267 "dma_device_id": "system", 00:06:04.267 "dma_device_type": 1 00:06:04.267 }, 00:06:04.267 { 00:06:04.267 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.267 "dma_device_type": 2 00:06:04.267 } 00:06:04.267 ], 00:06:04.267 "driver_specific": {} 00:06:04.267 } 00:06:04.267 ]' 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:06:04.267 ************************************ 00:06:04.267 END TEST rpc_plugins 00:06:04.267 ************************************ 00:06:04.267 10:55:09 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:06:04.267 00:06:04.267 real 0m0.175s 00:06:04.267 user 0m0.116s 00:06:04.267 sys 0m0.016s 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:04.267 10:55:09 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:06:04.267 10:55:09 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:06:04.267 10:55:09 rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:04.267 10:55:09 rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:04.267 10:55:09 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:04.267 ************************************ 00:06:04.267 START TEST rpc_trace_cmd_test 00:06:04.267 ************************************ 00:06:04.267 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1127 -- # rpc_trace_cmd_test 00:06:04.267 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:06:04.267 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:06:04.267 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.267 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:06:04.525 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid69264", 00:06:04.525 "tpoint_group_mask": "0x8", 00:06:04.525 "iscsi_conn": { 00:06:04.525 "mask": "0x2", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "scsi": { 00:06:04.525 "mask": "0x4", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "bdev": { 00:06:04.525 "mask": "0x8", 00:06:04.525 "tpoint_mask": "0xffffffffffffffff" 00:06:04.525 }, 00:06:04.525 "nvmf_rdma": { 00:06:04.525 "mask": "0x10", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "nvmf_tcp": { 00:06:04.525 "mask": "0x20", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "ftl": { 00:06:04.525 "mask": "0x40", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "blobfs": { 00:06:04.525 "mask": "0x80", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "dsa": { 00:06:04.525 "mask": "0x200", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "thread": { 00:06:04.525 "mask": "0x400", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "nvme_pcie": { 00:06:04.525 "mask": "0x800", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "iaa": { 00:06:04.525 "mask": "0x1000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "nvme_tcp": { 00:06:04.525 "mask": "0x2000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "bdev_nvme": { 00:06:04.525 "mask": "0x4000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "sock": { 00:06:04.525 "mask": "0x8000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "blob": { 00:06:04.525 "mask": "0x10000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "bdev_raid": { 00:06:04.525 "mask": "0x20000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 }, 00:06:04.525 "scheduler": { 00:06:04.525 "mask": "0x40000", 00:06:04.525 "tpoint_mask": "0x0" 00:06:04.525 } 00:06:04.525 }' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:06:04.525 ************************************ 00:06:04.525 END TEST rpc_trace_cmd_test 00:06:04.525 ************************************ 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:06:04.525 00:06:04.525 real 0m0.234s 00:06:04.525 user 0m0.182s 00:06:04.525 sys 0m0.042s 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:04.525 10:55:09 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:06:04.785 10:55:10 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:06:04.785 10:55:10 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:06:04.785 10:55:10 rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:04.785 10:55:10 rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:04.785 10:55:10 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 ************************************ 00:06:04.785 START TEST rpc_daemon_integrity 00:06:04.785 ************************************ 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1127 -- # rpc_integrity 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:06:04.785 { 00:06:04.785 "name": "Malloc2", 00:06:04.785 "aliases": [ 00:06:04.785 "d5c006cf-75d7-44d5-a952-f7140dd3c3ce" 00:06:04.785 ], 00:06:04.785 "product_name": "Malloc disk", 00:06:04.785 "block_size": 512, 00:06:04.785 "num_blocks": 16384, 00:06:04.785 "uuid": "d5c006cf-75d7-44d5-a952-f7140dd3c3ce", 00:06:04.785 "assigned_rate_limits": { 00:06:04.785 "rw_ios_per_sec": 0, 00:06:04.785 "rw_mbytes_per_sec": 0, 00:06:04.785 "r_mbytes_per_sec": 0, 00:06:04.785 "w_mbytes_per_sec": 0 00:06:04.785 }, 00:06:04.785 "claimed": false, 00:06:04.785 "zoned": false, 00:06:04.785 "supported_io_types": { 00:06:04.785 "read": true, 00:06:04.785 "write": true, 00:06:04.785 "unmap": true, 00:06:04.785 "flush": true, 00:06:04.785 "reset": true, 00:06:04.785 "nvme_admin": false, 00:06:04.785 "nvme_io": false, 00:06:04.785 "nvme_io_md": false, 00:06:04.785 "write_zeroes": true, 00:06:04.785 "zcopy": true, 00:06:04.785 "get_zone_info": false, 00:06:04.785 "zone_management": false, 00:06:04.785 "zone_append": false, 00:06:04.785 "compare": false, 00:06:04.785 "compare_and_write": false, 00:06:04.785 "abort": true, 00:06:04.785 "seek_hole": false, 00:06:04.785 "seek_data": false, 00:06:04.785 "copy": true, 00:06:04.785 "nvme_iov_md": false 00:06:04.785 }, 00:06:04.785 "memory_domains": [ 00:06:04.785 { 00:06:04.785 "dma_device_id": "system", 00:06:04.785 "dma_device_type": 1 00:06:04.785 }, 00:06:04.785 { 00:06:04.785 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.785 "dma_device_type": 2 00:06:04.785 } 00:06:04.785 ], 00:06:04.785 "driver_specific": {} 00:06:04.785 } 00:06:04.785 ]' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 [2024-10-29 10:55:10.188167] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:06:04.785 [2024-10-29 10:55:10.188237] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:04.785 [2024-10-29 10:55:10.188264] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:06:04.785 [2024-10-29 10:55:10.188274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:04.785 [2024-10-29 10:55:10.190800] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:04.785 [2024-10-29 10:55:10.190840] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:06:04.785 Passthru0 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:06:04.785 { 00:06:04.785 "name": "Malloc2", 00:06:04.785 "aliases": [ 00:06:04.785 "d5c006cf-75d7-44d5-a952-f7140dd3c3ce" 00:06:04.785 ], 00:06:04.785 "product_name": "Malloc disk", 00:06:04.785 "block_size": 512, 00:06:04.785 "num_blocks": 16384, 00:06:04.785 "uuid": "d5c006cf-75d7-44d5-a952-f7140dd3c3ce", 00:06:04.785 "assigned_rate_limits": { 00:06:04.785 "rw_ios_per_sec": 0, 00:06:04.785 "rw_mbytes_per_sec": 0, 00:06:04.785 "r_mbytes_per_sec": 0, 00:06:04.785 "w_mbytes_per_sec": 0 00:06:04.785 }, 00:06:04.785 "claimed": true, 00:06:04.785 "claim_type": "exclusive_write", 00:06:04.785 "zoned": false, 00:06:04.785 "supported_io_types": { 00:06:04.785 "read": true, 00:06:04.785 "write": true, 00:06:04.785 "unmap": true, 00:06:04.785 "flush": true, 00:06:04.785 "reset": true, 00:06:04.785 "nvme_admin": false, 00:06:04.785 "nvme_io": false, 00:06:04.785 "nvme_io_md": false, 00:06:04.785 "write_zeroes": true, 00:06:04.785 "zcopy": true, 00:06:04.785 "get_zone_info": false, 00:06:04.785 "zone_management": false, 00:06:04.785 "zone_append": false, 00:06:04.785 "compare": false, 00:06:04.785 "compare_and_write": false, 00:06:04.785 "abort": true, 00:06:04.785 "seek_hole": false, 00:06:04.785 "seek_data": false, 00:06:04.785 "copy": true, 00:06:04.785 "nvme_iov_md": false 00:06:04.785 }, 00:06:04.785 "memory_domains": [ 00:06:04.785 { 00:06:04.785 "dma_device_id": "system", 00:06:04.785 "dma_device_type": 1 00:06:04.785 }, 00:06:04.785 { 00:06:04.785 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.785 "dma_device_type": 2 00:06:04.785 } 00:06:04.785 ], 00:06:04.785 "driver_specific": {} 00:06:04.785 }, 00:06:04.785 { 00:06:04.785 "name": "Passthru0", 00:06:04.785 "aliases": [ 00:06:04.785 "4628d3ee-2c62-58c0-93ab-e610bc0b08bb" 00:06:04.785 ], 00:06:04.785 "product_name": "passthru", 00:06:04.785 "block_size": 512, 00:06:04.785 "num_blocks": 16384, 00:06:04.785 "uuid": "4628d3ee-2c62-58c0-93ab-e610bc0b08bb", 00:06:04.785 "assigned_rate_limits": { 00:06:04.785 "rw_ios_per_sec": 0, 00:06:04.785 "rw_mbytes_per_sec": 0, 00:06:04.785 "r_mbytes_per_sec": 0, 00:06:04.785 "w_mbytes_per_sec": 0 00:06:04.785 }, 00:06:04.785 "claimed": false, 00:06:04.785 "zoned": false, 00:06:04.785 "supported_io_types": { 00:06:04.785 "read": true, 00:06:04.785 "write": true, 00:06:04.785 "unmap": true, 00:06:04.785 "flush": true, 00:06:04.785 "reset": true, 00:06:04.785 "nvme_admin": false, 00:06:04.785 "nvme_io": false, 00:06:04.785 "nvme_io_md": false, 00:06:04.785 "write_zeroes": true, 00:06:04.785 "zcopy": true, 00:06:04.785 "get_zone_info": false, 00:06:04.785 "zone_management": false, 00:06:04.785 "zone_append": false, 00:06:04.785 "compare": false, 00:06:04.785 "compare_and_write": false, 00:06:04.785 "abort": true, 00:06:04.785 "seek_hole": false, 00:06:04.785 "seek_data": false, 00:06:04.785 "copy": true, 00:06:04.785 "nvme_iov_md": false 00:06:04.785 }, 00:06:04.785 "memory_domains": [ 00:06:04.785 { 00:06:04.785 "dma_device_id": "system", 00:06:04.785 "dma_device_type": 1 00:06:04.785 }, 00:06:04.785 { 00:06:04.785 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:04.785 "dma_device_type": 2 00:06:04.785 } 00:06:04.785 ], 00:06:04.785 "driver_specific": { 00:06:04.785 "passthru": { 00:06:04.785 "name": "Passthru0", 00:06:04.785 "base_bdev_name": "Malloc2" 00:06:04.785 } 00:06:04.785 } 00:06:04.785 } 00:06:04.785 ]' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:04.785 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:06:05.046 00:06:05.046 real 0m0.315s 00:06:05.046 user 0m0.175s 00:06:05.046 sys 0m0.067s 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:05.046 10:55:10 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:06:05.046 ************************************ 00:06:05.046 END TEST rpc_daemon_integrity 00:06:05.046 ************************************ 00:06:05.046 10:55:10 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:06:05.046 10:55:10 rpc -- rpc/rpc.sh@84 -- # killprocess 69264 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@952 -- # '[' -z 69264 ']' 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@956 -- # kill -0 69264 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@957 -- # uname 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69264 00:06:05.046 killing process with pid 69264 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69264' 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@971 -- # kill 69264 00:06:05.046 10:55:10 rpc -- common/autotest_common.sh@976 -- # wait 69264 00:06:05.616 00:06:05.616 real 0m3.083s 00:06:05.616 user 0m3.596s 00:06:05.616 sys 0m0.906s 00:06:05.616 10:55:11 rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:05.616 ************************************ 00:06:05.616 END TEST rpc 00:06:05.616 ************************************ 00:06:05.616 10:55:11 rpc -- common/autotest_common.sh@10 -- # set +x 00:06:05.877 10:55:11 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:05.877 10:55:11 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:05.877 10:55:11 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:05.877 10:55:11 -- common/autotest_common.sh@10 -- # set +x 00:06:05.877 ************************************ 00:06:05.877 START TEST skip_rpc 00:06:05.877 ************************************ 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:06:05.877 * Looking for test storage... 00:06:05.877 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@345 -- # : 1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:05.877 10:55:11 skip_rpc -- scripts/common.sh@368 -- # return 0 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:05.877 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.877 --rc genhtml_branch_coverage=1 00:06:05.877 --rc genhtml_function_coverage=1 00:06:05.877 --rc genhtml_legend=1 00:06:05.877 --rc geninfo_all_blocks=1 00:06:05.877 --rc geninfo_unexecuted_blocks=1 00:06:05.877 00:06:05.877 ' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:05.877 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.877 --rc genhtml_branch_coverage=1 00:06:05.877 --rc genhtml_function_coverage=1 00:06:05.877 --rc genhtml_legend=1 00:06:05.877 --rc geninfo_all_blocks=1 00:06:05.877 --rc geninfo_unexecuted_blocks=1 00:06:05.877 00:06:05.877 ' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:05.877 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.877 --rc genhtml_branch_coverage=1 00:06:05.877 --rc genhtml_function_coverage=1 00:06:05.877 --rc genhtml_legend=1 00:06:05.877 --rc geninfo_all_blocks=1 00:06:05.877 --rc geninfo_unexecuted_blocks=1 00:06:05.877 00:06:05.877 ' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:05.877 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.877 --rc genhtml_branch_coverage=1 00:06:05.877 --rc genhtml_function_coverage=1 00:06:05.877 --rc genhtml_legend=1 00:06:05.877 --rc geninfo_all_blocks=1 00:06:05.877 --rc geninfo_unexecuted_blocks=1 00:06:05.877 00:06:05.877 ' 00:06:05.877 10:55:11 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:05.877 10:55:11 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:05.877 10:55:11 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:05.877 10:55:11 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:05.877 ************************************ 00:06:05.877 START TEST skip_rpc 00:06:05.877 ************************************ 00:06:05.877 10:55:11 skip_rpc.skip_rpc -- common/autotest_common.sh@1127 -- # test_skip_rpc 00:06:05.877 10:55:11 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69472 00:06:05.877 10:55:11 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:05.877 10:55:11 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:06:05.877 10:55:11 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:06:06.137 [2024-10-29 10:55:11.461456] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:06.137 [2024-10-29 10:55:11.461589] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69472 ] 00:06:06.137 [2024-10-29 10:55:11.629957] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:06.397 [2024-10-29 10:55:11.655825] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69472 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@952 -- # '[' -z 69472 ']' 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # kill -0 69472 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@957 -- # uname 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69472 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:11.677 killing process with pid 69472 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69472' 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@971 -- # kill 69472 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@976 -- # wait 69472 00:06:11.677 00:06:11.677 real 0m5.421s 00:06:11.677 user 0m5.031s 00:06:11.677 sys 0m0.308s 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:11.677 10:55:16 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:11.677 ************************************ 00:06:11.677 END TEST skip_rpc 00:06:11.677 ************************************ 00:06:11.677 10:55:16 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:06:11.678 10:55:16 skip_rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:11.678 10:55:16 skip_rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:11.678 10:55:16 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:11.678 ************************************ 00:06:11.678 START TEST skip_rpc_with_json 00:06:11.678 ************************************ 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1127 -- # test_skip_rpc_with_json 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69554 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69554 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@833 -- # '[' -z 69554 ']' 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:11.678 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:11.678 10:55:16 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:11.678 [2024-10-29 10:55:16.942114] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:11.678 [2024-10-29 10:55:16.942267] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69554 ] 00:06:11.678 [2024-10-29 10:55:17.110834] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:11.678 [2024-10-29 10:55:17.136794] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@866 -- # return 0 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:12.638 [2024-10-29 10:55:17.764685] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:06:12.638 request: 00:06:12.638 { 00:06:12.638 "trtype": "tcp", 00:06:12.638 "method": "nvmf_get_transports", 00:06:12.638 "req_id": 1 00:06:12.638 } 00:06:12.638 Got JSON-RPC error response 00:06:12.638 response: 00:06:12.638 { 00:06:12.638 "code": -19, 00:06:12.638 "message": "No such device" 00:06:12.638 } 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:12.638 [2024-10-29 10:55:17.776791] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:12.638 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:12.638 { 00:06:12.638 "subsystems": [ 00:06:12.638 { 00:06:12.638 "subsystem": "fsdev", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "fsdev_set_opts", 00:06:12.638 "params": { 00:06:12.638 "fsdev_io_pool_size": 65535, 00:06:12.638 "fsdev_io_cache_size": 256 00:06:12.638 } 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "keyring", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "iobuf", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "iobuf_set_options", 00:06:12.638 "params": { 00:06:12.638 "small_pool_count": 8192, 00:06:12.638 "large_pool_count": 1024, 00:06:12.638 "small_bufsize": 8192, 00:06:12.638 "large_bufsize": 135168, 00:06:12.638 "enable_numa": false 00:06:12.638 } 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "sock", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "sock_set_default_impl", 00:06:12.638 "params": { 00:06:12.638 "impl_name": "posix" 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "sock_impl_set_options", 00:06:12.638 "params": { 00:06:12.638 "impl_name": "ssl", 00:06:12.638 "recv_buf_size": 4096, 00:06:12.638 "send_buf_size": 4096, 00:06:12.638 "enable_recv_pipe": true, 00:06:12.638 "enable_quickack": false, 00:06:12.638 "enable_placement_id": 0, 00:06:12.638 "enable_zerocopy_send_server": true, 00:06:12.638 "enable_zerocopy_send_client": false, 00:06:12.638 "zerocopy_threshold": 0, 00:06:12.638 "tls_version": 0, 00:06:12.638 "enable_ktls": false 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "sock_impl_set_options", 00:06:12.638 "params": { 00:06:12.638 "impl_name": "posix", 00:06:12.638 "recv_buf_size": 2097152, 00:06:12.638 "send_buf_size": 2097152, 00:06:12.638 "enable_recv_pipe": true, 00:06:12.638 "enable_quickack": false, 00:06:12.638 "enable_placement_id": 0, 00:06:12.638 "enable_zerocopy_send_server": true, 00:06:12.638 "enable_zerocopy_send_client": false, 00:06:12.638 "zerocopy_threshold": 0, 00:06:12.638 "tls_version": 0, 00:06:12.638 "enable_ktls": false 00:06:12.638 } 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "vmd", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "accel", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "accel_set_options", 00:06:12.638 "params": { 00:06:12.638 "small_cache_size": 128, 00:06:12.638 "large_cache_size": 16, 00:06:12.638 "task_count": 2048, 00:06:12.638 "sequence_count": 2048, 00:06:12.638 "buf_count": 2048 00:06:12.638 } 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "bdev", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "bdev_set_options", 00:06:12.638 "params": { 00:06:12.638 "bdev_io_pool_size": 65535, 00:06:12.638 "bdev_io_cache_size": 256, 00:06:12.638 "bdev_auto_examine": true, 00:06:12.638 "iobuf_small_cache_size": 128, 00:06:12.638 "iobuf_large_cache_size": 16 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "bdev_raid_set_options", 00:06:12.638 "params": { 00:06:12.638 "process_window_size_kb": 1024, 00:06:12.638 "process_max_bandwidth_mb_sec": 0 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "bdev_iscsi_set_options", 00:06:12.638 "params": { 00:06:12.638 "timeout_sec": 30 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "bdev_nvme_set_options", 00:06:12.638 "params": { 00:06:12.638 "action_on_timeout": "none", 00:06:12.638 "timeout_us": 0, 00:06:12.638 "timeout_admin_us": 0, 00:06:12.638 "keep_alive_timeout_ms": 10000, 00:06:12.638 "arbitration_burst": 0, 00:06:12.638 "low_priority_weight": 0, 00:06:12.638 "medium_priority_weight": 0, 00:06:12.638 "high_priority_weight": 0, 00:06:12.638 "nvme_adminq_poll_period_us": 10000, 00:06:12.638 "nvme_ioq_poll_period_us": 0, 00:06:12.638 "io_queue_requests": 0, 00:06:12.638 "delay_cmd_submit": true, 00:06:12.638 "transport_retry_count": 4, 00:06:12.638 "bdev_retry_count": 3, 00:06:12.638 "transport_ack_timeout": 0, 00:06:12.638 "ctrlr_loss_timeout_sec": 0, 00:06:12.638 "reconnect_delay_sec": 0, 00:06:12.638 "fast_io_fail_timeout_sec": 0, 00:06:12.638 "disable_auto_failback": false, 00:06:12.638 "generate_uuids": false, 00:06:12.638 "transport_tos": 0, 00:06:12.638 "nvme_error_stat": false, 00:06:12.638 "rdma_srq_size": 0, 00:06:12.638 "io_path_stat": false, 00:06:12.638 "allow_accel_sequence": false, 00:06:12.638 "rdma_max_cq_size": 0, 00:06:12.638 "rdma_cm_event_timeout_ms": 0, 00:06:12.638 "dhchap_digests": [ 00:06:12.638 "sha256", 00:06:12.638 "sha384", 00:06:12.638 "sha512" 00:06:12.638 ], 00:06:12.638 "dhchap_dhgroups": [ 00:06:12.638 "null", 00:06:12.638 "ffdhe2048", 00:06:12.638 "ffdhe3072", 00:06:12.638 "ffdhe4096", 00:06:12.638 "ffdhe6144", 00:06:12.638 "ffdhe8192" 00:06:12.638 ] 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "bdev_nvme_set_hotplug", 00:06:12.638 "params": { 00:06:12.638 "period_us": 100000, 00:06:12.638 "enable": false 00:06:12.638 } 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "method": "bdev_wait_for_examine" 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "scsi", 00:06:12.638 "config": null 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "scheduler", 00:06:12.638 "config": [ 00:06:12.638 { 00:06:12.638 "method": "framework_set_scheduler", 00:06:12.638 "params": { 00:06:12.638 "name": "static" 00:06:12.638 } 00:06:12.638 } 00:06:12.638 ] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "vhost_scsi", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "vhost_blk", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "ublk", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "nbd", 00:06:12.638 "config": [] 00:06:12.638 }, 00:06:12.638 { 00:06:12.638 "subsystem": "nvmf", 00:06:12.639 "config": [ 00:06:12.639 { 00:06:12.639 "method": "nvmf_set_config", 00:06:12.639 "params": { 00:06:12.639 "discovery_filter": "match_any", 00:06:12.639 "admin_cmd_passthru": { 00:06:12.639 "identify_ctrlr": false 00:06:12.639 }, 00:06:12.639 "dhchap_digests": [ 00:06:12.639 "sha256", 00:06:12.639 "sha384", 00:06:12.639 "sha512" 00:06:12.639 ], 00:06:12.639 "dhchap_dhgroups": [ 00:06:12.639 "null", 00:06:12.639 "ffdhe2048", 00:06:12.639 "ffdhe3072", 00:06:12.639 "ffdhe4096", 00:06:12.639 "ffdhe6144", 00:06:12.639 "ffdhe8192" 00:06:12.639 ] 00:06:12.639 } 00:06:12.639 }, 00:06:12.639 { 00:06:12.639 "method": "nvmf_set_max_subsystems", 00:06:12.639 "params": { 00:06:12.639 "max_subsystems": 1024 00:06:12.639 } 00:06:12.639 }, 00:06:12.639 { 00:06:12.639 "method": "nvmf_set_crdt", 00:06:12.639 "params": { 00:06:12.639 "crdt1": 0, 00:06:12.639 "crdt2": 0, 00:06:12.639 "crdt3": 0 00:06:12.639 } 00:06:12.639 }, 00:06:12.639 { 00:06:12.639 "method": "nvmf_create_transport", 00:06:12.639 "params": { 00:06:12.639 "trtype": "TCP", 00:06:12.639 "max_queue_depth": 128, 00:06:12.639 "max_io_qpairs_per_ctrlr": 127, 00:06:12.639 "in_capsule_data_size": 4096, 00:06:12.639 "max_io_size": 131072, 00:06:12.639 "io_unit_size": 131072, 00:06:12.639 "max_aq_depth": 128, 00:06:12.639 "num_shared_buffers": 511, 00:06:12.639 "buf_cache_size": 4294967295, 00:06:12.639 "dif_insert_or_strip": false, 00:06:12.639 "zcopy": false, 00:06:12.639 "c2h_success": true, 00:06:12.639 "sock_priority": 0, 00:06:12.639 "abort_timeout_sec": 1, 00:06:12.639 "ack_timeout": 0, 00:06:12.639 "data_wr_pool_size": 0 00:06:12.639 } 00:06:12.639 } 00:06:12.639 ] 00:06:12.639 }, 00:06:12.639 { 00:06:12.639 "subsystem": "iscsi", 00:06:12.639 "config": [ 00:06:12.639 { 00:06:12.639 "method": "iscsi_set_options", 00:06:12.639 "params": { 00:06:12.639 "node_base": "iqn.2016-06.io.spdk", 00:06:12.639 "max_sessions": 128, 00:06:12.639 "max_connections_per_session": 2, 00:06:12.639 "max_queue_depth": 64, 00:06:12.639 "default_time2wait": 2, 00:06:12.639 "default_time2retain": 20, 00:06:12.639 "first_burst_length": 8192, 00:06:12.639 "immediate_data": true, 00:06:12.639 "allow_duplicated_isid": false, 00:06:12.639 "error_recovery_level": 0, 00:06:12.639 "nop_timeout": 60, 00:06:12.639 "nop_in_interval": 30, 00:06:12.639 "disable_chap": false, 00:06:12.639 "require_chap": false, 00:06:12.639 "mutual_chap": false, 00:06:12.639 "chap_group": 0, 00:06:12.639 "max_large_datain_per_connection": 64, 00:06:12.639 "max_r2t_per_connection": 4, 00:06:12.639 "pdu_pool_size": 36864, 00:06:12.639 "immediate_data_pool_size": 16384, 00:06:12.639 "data_out_pool_size": 2048 00:06:12.639 } 00:06:12.639 } 00:06:12.639 ] 00:06:12.639 } 00:06:12.639 ] 00:06:12.639 } 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69554 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # '[' -z 69554 ']' 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # kill -0 69554 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@957 -- # uname 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69554 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:12.639 killing process with pid 69554 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69554' 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@971 -- # kill 69554 00:06:12.639 10:55:17 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@976 -- # wait 69554 00:06:12.899 10:55:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69583 00:06:12.899 10:55:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:12.899 10:55:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69583 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # '[' -z 69583 ']' 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # kill -0 69583 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@957 -- # uname 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69583 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:18.179 killing process with pid 69583 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69583' 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@971 -- # kill 69583 00:06:18.179 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@976 -- # wait 69583 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:18.439 00:06:18.439 real 0m6.930s 00:06:18.439 user 0m6.497s 00:06:18.439 sys 0m0.710s 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:18.439 ************************************ 00:06:18.439 END TEST skip_rpc_with_json 00:06:18.439 ************************************ 00:06:18.439 10:55:23 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:18.439 10:55:23 skip_rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:18.439 10:55:23 skip_rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:18.439 10:55:23 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:18.439 ************************************ 00:06:18.439 START TEST skip_rpc_with_delay 00:06:18.439 ************************************ 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1127 -- # test_skip_rpc_with_delay 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:18.439 10:55:23 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:18.699 [2024-10-29 10:55:23.942253] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:18.699 00:06:18.699 real 0m0.163s 00:06:18.699 user 0m0.097s 00:06:18.699 sys 0m0.065s 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:18.699 10:55:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:18.699 ************************************ 00:06:18.699 END TEST skip_rpc_with_delay 00:06:18.699 ************************************ 00:06:18.699 10:55:24 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:18.699 10:55:24 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:18.699 10:55:24 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:18.700 10:55:24 skip_rpc -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:18.700 10:55:24 skip_rpc -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:18.700 10:55:24 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:18.700 ************************************ 00:06:18.700 START TEST exit_on_failed_rpc_init 00:06:18.700 ************************************ 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1127 -- # test_exit_on_failed_rpc_init 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69694 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69694 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@833 -- # '[' -z 69694 ']' 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:18.700 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:18.700 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:18.700 [2024-10-29 10:55:24.176195] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:18.700 [2024-10-29 10:55:24.176314] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69694 ] 00:06:18.959 [2024-10-29 10:55:24.345237] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:18.959 [2024-10-29 10:55:24.373721] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@866 -- # return 0 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:19.528 10:55:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:19.788 [2024-10-29 10:55:25.095891] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:19.788 [2024-10-29 10:55:25.096051] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69711 ] 00:06:19.788 [2024-10-29 10:55:25.267380] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:20.048 [2024-10-29 10:55:25.293588] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:20.048 [2024-10-29 10:55:25.293672] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:20.048 [2024-10-29 10:55:25.293689] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:20.048 [2024-10-29 10:55:25.293703] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69694 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@952 -- # '[' -z 69694 ']' 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # kill -0 69694 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@957 -- # uname 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69694 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:20.048 killing process with pid 69694 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69694' 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@971 -- # kill 69694 00:06:20.048 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@976 -- # wait 69694 00:06:20.308 00:06:20.308 real 0m1.712s 00:06:20.308 user 0m1.834s 00:06:20.308 sys 0m0.499s 00:06:20.308 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:20.308 10:55:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:20.308 ************************************ 00:06:20.308 END TEST exit_on_failed_rpc_init 00:06:20.308 ************************************ 00:06:20.568 10:55:25 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:20.568 00:06:20.568 real 0m14.723s 00:06:20.568 user 0m13.669s 00:06:20.568 sys 0m1.880s 00:06:20.568 10:55:25 skip_rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:20.568 10:55:25 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:20.568 ************************************ 00:06:20.568 END TEST skip_rpc 00:06:20.568 ************************************ 00:06:20.568 10:55:25 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:20.568 10:55:25 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:20.568 10:55:25 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:20.568 10:55:25 -- common/autotest_common.sh@10 -- # set +x 00:06:20.568 ************************************ 00:06:20.568 START TEST rpc_client 00:06:20.568 ************************************ 00:06:20.568 10:55:25 rpc_client -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:20.568 * Looking for test storage... 00:06:20.568 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:20.568 10:55:26 rpc_client -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:20.568 10:55:26 rpc_client -- common/autotest_common.sh@1691 -- # lcov --version 00:06:20.568 10:55:26 rpc_client -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:20.829 10:55:26 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:20.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.829 --rc genhtml_branch_coverage=1 00:06:20.829 --rc genhtml_function_coverage=1 00:06:20.829 --rc genhtml_legend=1 00:06:20.829 --rc geninfo_all_blocks=1 00:06:20.829 --rc geninfo_unexecuted_blocks=1 00:06:20.829 00:06:20.829 ' 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:20.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.829 --rc genhtml_branch_coverage=1 00:06:20.829 --rc genhtml_function_coverage=1 00:06:20.829 --rc genhtml_legend=1 00:06:20.829 --rc geninfo_all_blocks=1 00:06:20.829 --rc geninfo_unexecuted_blocks=1 00:06:20.829 00:06:20.829 ' 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:20.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.829 --rc genhtml_branch_coverage=1 00:06:20.829 --rc genhtml_function_coverage=1 00:06:20.829 --rc genhtml_legend=1 00:06:20.829 --rc geninfo_all_blocks=1 00:06:20.829 --rc geninfo_unexecuted_blocks=1 00:06:20.829 00:06:20.829 ' 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:20.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:20.829 --rc genhtml_branch_coverage=1 00:06:20.829 --rc genhtml_function_coverage=1 00:06:20.829 --rc genhtml_legend=1 00:06:20.829 --rc geninfo_all_blocks=1 00:06:20.829 --rc geninfo_unexecuted_blocks=1 00:06:20.829 00:06:20.829 ' 00:06:20.829 10:55:26 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:20.829 OK 00:06:20.829 10:55:26 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:20.829 00:06:20.829 real 0m0.295s 00:06:20.829 user 0m0.152s 00:06:20.829 sys 0m0.162s 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:20.829 10:55:26 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:20.829 ************************************ 00:06:20.829 END TEST rpc_client 00:06:20.829 ************************************ 00:06:20.829 10:55:26 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:20.829 10:55:26 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:20.829 10:55:26 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:20.829 10:55:26 -- common/autotest_common.sh@10 -- # set +x 00:06:20.829 ************************************ 00:06:20.829 START TEST json_config 00:06:20.829 ************************************ 00:06:20.829 10:55:26 json_config -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1691 -- # lcov --version 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:21.090 10:55:26 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:21.090 10:55:26 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:21.090 10:55:26 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:21.090 10:55:26 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:21.090 10:55:26 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:21.090 10:55:26 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:21.090 10:55:26 json_config -- scripts/common.sh@345 -- # : 1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:21.090 10:55:26 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:21.090 10:55:26 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@353 -- # local d=1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:21.090 10:55:26 json_config -- scripts/common.sh@355 -- # echo 1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:21.090 10:55:26 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@353 -- # local d=2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:21.090 10:55:26 json_config -- scripts/common.sh@355 -- # echo 2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:21.090 10:55:26 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:21.090 10:55:26 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:21.090 10:55:26 json_config -- scripts/common.sh@368 -- # return 0 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:21.090 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.090 --rc genhtml_branch_coverage=1 00:06:21.090 --rc genhtml_function_coverage=1 00:06:21.090 --rc genhtml_legend=1 00:06:21.090 --rc geninfo_all_blocks=1 00:06:21.090 --rc geninfo_unexecuted_blocks=1 00:06:21.090 00:06:21.090 ' 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:21.090 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.090 --rc genhtml_branch_coverage=1 00:06:21.090 --rc genhtml_function_coverage=1 00:06:21.090 --rc genhtml_legend=1 00:06:21.090 --rc geninfo_all_blocks=1 00:06:21.090 --rc geninfo_unexecuted_blocks=1 00:06:21.090 00:06:21.090 ' 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:21.090 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.090 --rc genhtml_branch_coverage=1 00:06:21.090 --rc genhtml_function_coverage=1 00:06:21.090 --rc genhtml_legend=1 00:06:21.090 --rc geninfo_all_blocks=1 00:06:21.090 --rc geninfo_unexecuted_blocks=1 00:06:21.090 00:06:21.090 ' 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:21.090 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.090 --rc genhtml_branch_coverage=1 00:06:21.090 --rc genhtml_function_coverage=1 00:06:21.090 --rc genhtml_legend=1 00:06:21.090 --rc geninfo_all_blocks=1 00:06:21.090 --rc geninfo_unexecuted_blocks=1 00:06:21.090 00:06:21.090 ' 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:21.090 10:55:26 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:21.090 10:55:26 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:21.090 10:55:26 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:21.090 10:55:26 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:21.090 10:55:26 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.090 10:55:26 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.090 10:55:26 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.090 10:55:26 json_config -- paths/export.sh@5 -- # export PATH 00:06:21.090 10:55:26 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@51 -- # : 0 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:21.090 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:21.090 10:55:26 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:21.090 WARNING: No tests are enabled so not running JSON configuration tests 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:21.090 10:55:26 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:21.090 00:06:21.090 real 0m0.217s 00:06:21.090 user 0m0.133s 00:06:21.090 sys 0m0.095s 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:21.090 10:55:26 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:21.090 ************************************ 00:06:21.090 END TEST json_config 00:06:21.090 ************************************ 00:06:21.090 10:55:26 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:21.091 10:55:26 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:21.091 10:55:26 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:21.091 10:55:26 -- common/autotest_common.sh@10 -- # set +x 00:06:21.091 ************************************ 00:06:21.091 START TEST json_config_extra_key 00:06:21.091 ************************************ 00:06:21.091 10:55:26 json_config_extra_key -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1691 -- # lcov --version 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:21.351 10:55:26 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:21.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.351 --rc genhtml_branch_coverage=1 00:06:21.351 --rc genhtml_function_coverage=1 00:06:21.351 --rc genhtml_legend=1 00:06:21.351 --rc geninfo_all_blocks=1 00:06:21.351 --rc geninfo_unexecuted_blocks=1 00:06:21.351 00:06:21.351 ' 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:21.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.351 --rc genhtml_branch_coverage=1 00:06:21.351 --rc genhtml_function_coverage=1 00:06:21.351 --rc genhtml_legend=1 00:06:21.351 --rc geninfo_all_blocks=1 00:06:21.351 --rc geninfo_unexecuted_blocks=1 00:06:21.351 00:06:21.351 ' 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:21.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.351 --rc genhtml_branch_coverage=1 00:06:21.351 --rc genhtml_function_coverage=1 00:06:21.351 --rc genhtml_legend=1 00:06:21.351 --rc geninfo_all_blocks=1 00:06:21.351 --rc geninfo_unexecuted_blocks=1 00:06:21.351 00:06:21.351 ' 00:06:21.351 10:55:26 json_config_extra_key -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:21.351 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:21.351 --rc genhtml_branch_coverage=1 00:06:21.351 --rc genhtml_function_coverage=1 00:06:21.351 --rc genhtml_legend=1 00:06:21.351 --rc geninfo_all_blocks=1 00:06:21.351 --rc geninfo_unexecuted_blocks=1 00:06:21.351 00:06:21.351 ' 00:06:21.351 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:21.351 10:55:26 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:21.351 10:55:26 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:21.351 10:55:26 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=74b74ee9-5d23-4e0d-9a06-c0cff30226f0 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:21.352 10:55:26 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:21.352 10:55:26 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:21.352 10:55:26 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:21.352 10:55:26 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:21.352 10:55:26 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.352 10:55:26 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.352 10:55:26 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.352 10:55:26 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:21.352 10:55:26 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:21.352 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:21.352 10:55:26 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:21.352 INFO: launching applications... 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:21.352 10:55:26 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69895 00:06:21.352 Waiting for target to run... 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69895 /var/tmp/spdk_tgt.sock 00:06:21.352 10:55:26 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@833 -- # '[' -z 69895 ']' 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:21.352 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:21.352 10:55:26 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:21.612 [2024-10-29 10:55:26.866463] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:21.612 [2024-10-29 10:55:26.866612] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69895 ] 00:06:21.872 [2024-10-29 10:55:27.248874] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:21.872 [2024-10-29 10:55:27.266583] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.441 10:55:27 json_config_extra_key -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:22.441 10:55:27 json_config_extra_key -- common/autotest_common.sh@866 -- # return 0 00:06:22.441 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:06:22.441 INFO: shutting down applications... 00:06:22.441 10:55:27 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:06:22.441 10:55:27 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69895 ]] 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69895 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69895 00:06:22.441 10:55:27 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69895 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@43 -- # break 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:06:22.699 SPDK target shutdown done 00:06:22.699 10:55:28 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:06:22.699 Success 00:06:22.699 10:55:28 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:06:22.699 00:06:22.699 real 0m1.643s 00:06:22.699 user 0m1.363s 00:06:22.699 sys 0m0.457s 00:06:22.699 10:55:28 json_config_extra_key -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:22.699 10:55:28 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:22.699 ************************************ 00:06:22.699 END TEST json_config_extra_key 00:06:22.699 ************************************ 00:06:22.957 10:55:28 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:22.957 10:55:28 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:22.957 10:55:28 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:22.957 10:55:28 -- common/autotest_common.sh@10 -- # set +x 00:06:22.957 ************************************ 00:06:22.957 START TEST alias_rpc 00:06:22.957 ************************************ 00:06:22.957 10:55:28 alias_rpc -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:22.957 * Looking for test storage... 00:06:22.957 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:06:22.957 10:55:28 alias_rpc -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:22.958 10:55:28 alias_rpc -- common/autotest_common.sh@1691 -- # lcov --version 00:06:22.958 10:55:28 alias_rpc -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:22.958 10:55:28 alias_rpc -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@345 -- # : 1 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:22.958 10:55:28 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:23.215 10:55:28 alias_rpc -- scripts/common.sh@368 -- # return 0 00:06:23.215 10:55:28 alias_rpc -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:23.215 10:55:28 alias_rpc -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:23.215 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:23.215 --rc genhtml_branch_coverage=1 00:06:23.215 --rc genhtml_function_coverage=1 00:06:23.215 --rc genhtml_legend=1 00:06:23.215 --rc geninfo_all_blocks=1 00:06:23.215 --rc geninfo_unexecuted_blocks=1 00:06:23.215 00:06:23.215 ' 00:06:23.215 10:55:28 alias_rpc -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:23.215 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:23.215 --rc genhtml_branch_coverage=1 00:06:23.215 --rc genhtml_function_coverage=1 00:06:23.215 --rc genhtml_legend=1 00:06:23.215 --rc geninfo_all_blocks=1 00:06:23.215 --rc geninfo_unexecuted_blocks=1 00:06:23.215 00:06:23.215 ' 00:06:23.215 10:55:28 alias_rpc -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:23.215 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:23.215 --rc genhtml_branch_coverage=1 00:06:23.215 --rc genhtml_function_coverage=1 00:06:23.215 --rc genhtml_legend=1 00:06:23.215 --rc geninfo_all_blocks=1 00:06:23.215 --rc geninfo_unexecuted_blocks=1 00:06:23.215 00:06:23.215 ' 00:06:23.215 10:55:28 alias_rpc -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:23.215 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:23.215 --rc genhtml_branch_coverage=1 00:06:23.215 --rc genhtml_function_coverage=1 00:06:23.216 --rc genhtml_legend=1 00:06:23.216 --rc geninfo_all_blocks=1 00:06:23.216 --rc geninfo_unexecuted_blocks=1 00:06:23.216 00:06:23.216 ' 00:06:23.216 10:55:28 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:06:23.216 10:55:28 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69974 00:06:23.216 10:55:28 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69974 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@833 -- # '[' -z 69974 ']' 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:23.216 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:23.216 10:55:28 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:23.216 10:55:28 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:23.216 [2024-10-29 10:55:28.566350] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:23.216 [2024-10-29 10:55:28.566486] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69974 ] 00:06:23.475 [2024-10-29 10:55:28.720911] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:23.475 [2024-10-29 10:55:28.749937] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:24.042 10:55:29 alias_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:24.042 10:55:29 alias_rpc -- common/autotest_common.sh@866 -- # return 0 00:06:24.042 10:55:29 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:06:24.309 10:55:29 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69974 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@952 -- # '[' -z 69974 ']' 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@956 -- # kill -0 69974 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@957 -- # uname 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 69974 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:24.309 killing process with pid 69974 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@970 -- # echo 'killing process with pid 69974' 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@971 -- # kill 69974 00:06:24.309 10:55:29 alias_rpc -- common/autotest_common.sh@976 -- # wait 69974 00:06:24.568 00:06:24.568 real 0m1.745s 00:06:24.568 user 0m1.778s 00:06:24.568 sys 0m0.494s 00:06:24.568 10:55:30 alias_rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:24.568 10:55:30 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:24.568 ************************************ 00:06:24.568 END TEST alias_rpc 00:06:24.568 ************************************ 00:06:24.568 10:55:30 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:06:24.568 10:55:30 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:24.568 10:55:30 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:24.568 10:55:30 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:24.568 10:55:30 -- common/autotest_common.sh@10 -- # set +x 00:06:24.568 ************************************ 00:06:24.568 START TEST spdkcli_tcp 00:06:24.568 ************************************ 00:06:24.568 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:24.828 * Looking for test storage... 00:06:24.828 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1691 -- # lcov --version 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:24.828 10:55:30 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:24.828 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:24.828 --rc genhtml_branch_coverage=1 00:06:24.828 --rc genhtml_function_coverage=1 00:06:24.828 --rc genhtml_legend=1 00:06:24.828 --rc geninfo_all_blocks=1 00:06:24.828 --rc geninfo_unexecuted_blocks=1 00:06:24.828 00:06:24.828 ' 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:24.828 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:24.828 --rc genhtml_branch_coverage=1 00:06:24.828 --rc genhtml_function_coverage=1 00:06:24.828 --rc genhtml_legend=1 00:06:24.828 --rc geninfo_all_blocks=1 00:06:24.828 --rc geninfo_unexecuted_blocks=1 00:06:24.828 00:06:24.828 ' 00:06:24.828 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:24.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:24.829 --rc genhtml_branch_coverage=1 00:06:24.829 --rc genhtml_function_coverage=1 00:06:24.829 --rc genhtml_legend=1 00:06:24.829 --rc geninfo_all_blocks=1 00:06:24.829 --rc geninfo_unexecuted_blocks=1 00:06:24.829 00:06:24.829 ' 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:24.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:24.829 --rc genhtml_branch_coverage=1 00:06:24.829 --rc genhtml_function_coverage=1 00:06:24.829 --rc genhtml_legend=1 00:06:24.829 --rc geninfo_all_blocks=1 00:06:24.829 --rc geninfo_unexecuted_blocks=1 00:06:24.829 00:06:24.829 ' 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=70059 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:06:24.829 10:55:30 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 70059 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@833 -- # '[' -z 70059 ']' 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:24.829 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:24.829 10:55:30 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:25.089 [2024-10-29 10:55:30.387613] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:25.089 [2024-10-29 10:55:30.387774] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70059 ] 00:06:25.089 [2024-10-29 10:55:30.556608] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:25.089 [2024-10-29 10:55:30.587025] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:25.089 [2024-10-29 10:55:30.587125] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:26.029 10:55:31 spdkcli_tcp -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:26.029 10:55:31 spdkcli_tcp -- common/autotest_common.sh@866 -- # return 0 00:06:26.029 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:06:26.029 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=70065 00:06:26.029 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:06:26.029 [ 00:06:26.029 "bdev_malloc_delete", 00:06:26.029 "bdev_malloc_create", 00:06:26.029 "bdev_null_resize", 00:06:26.029 "bdev_null_delete", 00:06:26.029 "bdev_null_create", 00:06:26.029 "bdev_nvme_cuse_unregister", 00:06:26.029 "bdev_nvme_cuse_register", 00:06:26.029 "bdev_opal_new_user", 00:06:26.029 "bdev_opal_set_lock_state", 00:06:26.029 "bdev_opal_delete", 00:06:26.029 "bdev_opal_get_info", 00:06:26.029 "bdev_opal_create", 00:06:26.029 "bdev_nvme_opal_revert", 00:06:26.029 "bdev_nvme_opal_init", 00:06:26.029 "bdev_nvme_send_cmd", 00:06:26.029 "bdev_nvme_set_keys", 00:06:26.029 "bdev_nvme_get_path_iostat", 00:06:26.029 "bdev_nvme_get_mdns_discovery_info", 00:06:26.029 "bdev_nvme_stop_mdns_discovery", 00:06:26.029 "bdev_nvme_start_mdns_discovery", 00:06:26.029 "bdev_nvme_set_multipath_policy", 00:06:26.029 "bdev_nvme_set_preferred_path", 00:06:26.029 "bdev_nvme_get_io_paths", 00:06:26.029 "bdev_nvme_remove_error_injection", 00:06:26.029 "bdev_nvme_add_error_injection", 00:06:26.029 "bdev_nvme_get_discovery_info", 00:06:26.029 "bdev_nvme_stop_discovery", 00:06:26.029 "bdev_nvme_start_discovery", 00:06:26.029 "bdev_nvme_get_controller_health_info", 00:06:26.029 "bdev_nvme_disable_controller", 00:06:26.029 "bdev_nvme_enable_controller", 00:06:26.029 "bdev_nvme_reset_controller", 00:06:26.029 "bdev_nvme_get_transport_statistics", 00:06:26.029 "bdev_nvme_apply_firmware", 00:06:26.029 "bdev_nvme_detach_controller", 00:06:26.029 "bdev_nvme_get_controllers", 00:06:26.029 "bdev_nvme_attach_controller", 00:06:26.029 "bdev_nvme_set_hotplug", 00:06:26.029 "bdev_nvme_set_options", 00:06:26.029 "bdev_passthru_delete", 00:06:26.030 "bdev_passthru_create", 00:06:26.030 "bdev_lvol_set_parent_bdev", 00:06:26.030 "bdev_lvol_set_parent", 00:06:26.030 "bdev_lvol_check_shallow_copy", 00:06:26.030 "bdev_lvol_start_shallow_copy", 00:06:26.030 "bdev_lvol_grow_lvstore", 00:06:26.030 "bdev_lvol_get_lvols", 00:06:26.030 "bdev_lvol_get_lvstores", 00:06:26.030 "bdev_lvol_delete", 00:06:26.030 "bdev_lvol_set_read_only", 00:06:26.030 "bdev_lvol_resize", 00:06:26.030 "bdev_lvol_decouple_parent", 00:06:26.030 "bdev_lvol_inflate", 00:06:26.030 "bdev_lvol_rename", 00:06:26.030 "bdev_lvol_clone_bdev", 00:06:26.030 "bdev_lvol_clone", 00:06:26.030 "bdev_lvol_snapshot", 00:06:26.030 "bdev_lvol_create", 00:06:26.030 "bdev_lvol_delete_lvstore", 00:06:26.030 "bdev_lvol_rename_lvstore", 00:06:26.030 "bdev_lvol_create_lvstore", 00:06:26.030 "bdev_raid_set_options", 00:06:26.030 "bdev_raid_remove_base_bdev", 00:06:26.030 "bdev_raid_add_base_bdev", 00:06:26.030 "bdev_raid_delete", 00:06:26.030 "bdev_raid_create", 00:06:26.030 "bdev_raid_get_bdevs", 00:06:26.030 "bdev_error_inject_error", 00:06:26.030 "bdev_error_delete", 00:06:26.030 "bdev_error_create", 00:06:26.030 "bdev_split_delete", 00:06:26.030 "bdev_split_create", 00:06:26.030 "bdev_delay_delete", 00:06:26.030 "bdev_delay_create", 00:06:26.030 "bdev_delay_update_latency", 00:06:26.030 "bdev_zone_block_delete", 00:06:26.030 "bdev_zone_block_create", 00:06:26.030 "blobfs_create", 00:06:26.030 "blobfs_detect", 00:06:26.030 "blobfs_set_cache_size", 00:06:26.030 "bdev_aio_delete", 00:06:26.030 "bdev_aio_rescan", 00:06:26.030 "bdev_aio_create", 00:06:26.030 "bdev_ftl_set_property", 00:06:26.030 "bdev_ftl_get_properties", 00:06:26.030 "bdev_ftl_get_stats", 00:06:26.030 "bdev_ftl_unmap", 00:06:26.030 "bdev_ftl_unload", 00:06:26.030 "bdev_ftl_delete", 00:06:26.030 "bdev_ftl_load", 00:06:26.030 "bdev_ftl_create", 00:06:26.030 "bdev_virtio_attach_controller", 00:06:26.030 "bdev_virtio_scsi_get_devices", 00:06:26.030 "bdev_virtio_detach_controller", 00:06:26.030 "bdev_virtio_blk_set_hotplug", 00:06:26.030 "bdev_iscsi_delete", 00:06:26.030 "bdev_iscsi_create", 00:06:26.030 "bdev_iscsi_set_options", 00:06:26.030 "accel_error_inject_error", 00:06:26.030 "ioat_scan_accel_module", 00:06:26.030 "dsa_scan_accel_module", 00:06:26.030 "iaa_scan_accel_module", 00:06:26.030 "keyring_file_remove_key", 00:06:26.030 "keyring_file_add_key", 00:06:26.030 "keyring_linux_set_options", 00:06:26.030 "fsdev_aio_delete", 00:06:26.030 "fsdev_aio_create", 00:06:26.030 "iscsi_get_histogram", 00:06:26.030 "iscsi_enable_histogram", 00:06:26.030 "iscsi_set_options", 00:06:26.030 "iscsi_get_auth_groups", 00:06:26.030 "iscsi_auth_group_remove_secret", 00:06:26.030 "iscsi_auth_group_add_secret", 00:06:26.030 "iscsi_delete_auth_group", 00:06:26.030 "iscsi_create_auth_group", 00:06:26.030 "iscsi_set_discovery_auth", 00:06:26.030 "iscsi_get_options", 00:06:26.030 "iscsi_target_node_request_logout", 00:06:26.030 "iscsi_target_node_set_redirect", 00:06:26.030 "iscsi_target_node_set_auth", 00:06:26.030 "iscsi_target_node_add_lun", 00:06:26.030 "iscsi_get_stats", 00:06:26.030 "iscsi_get_connections", 00:06:26.030 "iscsi_portal_group_set_auth", 00:06:26.030 "iscsi_start_portal_group", 00:06:26.030 "iscsi_delete_portal_group", 00:06:26.030 "iscsi_create_portal_group", 00:06:26.030 "iscsi_get_portal_groups", 00:06:26.030 "iscsi_delete_target_node", 00:06:26.030 "iscsi_target_node_remove_pg_ig_maps", 00:06:26.030 "iscsi_target_node_add_pg_ig_maps", 00:06:26.030 "iscsi_create_target_node", 00:06:26.030 "iscsi_get_target_nodes", 00:06:26.030 "iscsi_delete_initiator_group", 00:06:26.030 "iscsi_initiator_group_remove_initiators", 00:06:26.030 "iscsi_initiator_group_add_initiators", 00:06:26.030 "iscsi_create_initiator_group", 00:06:26.030 "iscsi_get_initiator_groups", 00:06:26.030 "nvmf_set_crdt", 00:06:26.030 "nvmf_set_config", 00:06:26.030 "nvmf_set_max_subsystems", 00:06:26.030 "nvmf_stop_mdns_prr", 00:06:26.030 "nvmf_publish_mdns_prr", 00:06:26.030 "nvmf_subsystem_get_listeners", 00:06:26.030 "nvmf_subsystem_get_qpairs", 00:06:26.030 "nvmf_subsystem_get_controllers", 00:06:26.030 "nvmf_get_stats", 00:06:26.030 "nvmf_get_transports", 00:06:26.030 "nvmf_create_transport", 00:06:26.030 "nvmf_get_targets", 00:06:26.030 "nvmf_delete_target", 00:06:26.030 "nvmf_create_target", 00:06:26.030 "nvmf_subsystem_allow_any_host", 00:06:26.030 "nvmf_subsystem_set_keys", 00:06:26.030 "nvmf_subsystem_remove_host", 00:06:26.030 "nvmf_subsystem_add_host", 00:06:26.030 "nvmf_ns_remove_host", 00:06:26.030 "nvmf_ns_add_host", 00:06:26.030 "nvmf_subsystem_remove_ns", 00:06:26.030 "nvmf_subsystem_set_ns_ana_group", 00:06:26.030 "nvmf_subsystem_add_ns", 00:06:26.030 "nvmf_subsystem_listener_set_ana_state", 00:06:26.030 "nvmf_discovery_get_referrals", 00:06:26.030 "nvmf_discovery_remove_referral", 00:06:26.030 "nvmf_discovery_add_referral", 00:06:26.030 "nvmf_subsystem_remove_listener", 00:06:26.030 "nvmf_subsystem_add_listener", 00:06:26.030 "nvmf_delete_subsystem", 00:06:26.030 "nvmf_create_subsystem", 00:06:26.030 "nvmf_get_subsystems", 00:06:26.030 "env_dpdk_get_mem_stats", 00:06:26.030 "nbd_get_disks", 00:06:26.030 "nbd_stop_disk", 00:06:26.030 "nbd_start_disk", 00:06:26.030 "ublk_recover_disk", 00:06:26.030 "ublk_get_disks", 00:06:26.030 "ublk_stop_disk", 00:06:26.030 "ublk_start_disk", 00:06:26.030 "ublk_destroy_target", 00:06:26.030 "ublk_create_target", 00:06:26.030 "virtio_blk_create_transport", 00:06:26.030 "virtio_blk_get_transports", 00:06:26.030 "vhost_controller_set_coalescing", 00:06:26.030 "vhost_get_controllers", 00:06:26.030 "vhost_delete_controller", 00:06:26.030 "vhost_create_blk_controller", 00:06:26.030 "vhost_scsi_controller_remove_target", 00:06:26.030 "vhost_scsi_controller_add_target", 00:06:26.030 "vhost_start_scsi_controller", 00:06:26.030 "vhost_create_scsi_controller", 00:06:26.030 "thread_set_cpumask", 00:06:26.030 "scheduler_set_options", 00:06:26.030 "framework_get_governor", 00:06:26.030 "framework_get_scheduler", 00:06:26.030 "framework_set_scheduler", 00:06:26.030 "framework_get_reactors", 00:06:26.030 "thread_get_io_channels", 00:06:26.030 "thread_get_pollers", 00:06:26.030 "thread_get_stats", 00:06:26.030 "framework_monitor_context_switch", 00:06:26.030 "spdk_kill_instance", 00:06:26.030 "log_enable_timestamps", 00:06:26.030 "log_get_flags", 00:06:26.030 "log_clear_flag", 00:06:26.030 "log_set_flag", 00:06:26.030 "log_get_level", 00:06:26.030 "log_set_level", 00:06:26.030 "log_get_print_level", 00:06:26.030 "log_set_print_level", 00:06:26.030 "framework_enable_cpumask_locks", 00:06:26.030 "framework_disable_cpumask_locks", 00:06:26.030 "framework_wait_init", 00:06:26.030 "framework_start_init", 00:06:26.030 "scsi_get_devices", 00:06:26.030 "bdev_get_histogram", 00:06:26.030 "bdev_enable_histogram", 00:06:26.030 "bdev_set_qos_limit", 00:06:26.030 "bdev_set_qd_sampling_period", 00:06:26.030 "bdev_get_bdevs", 00:06:26.030 "bdev_reset_iostat", 00:06:26.030 "bdev_get_iostat", 00:06:26.030 "bdev_examine", 00:06:26.030 "bdev_wait_for_examine", 00:06:26.030 "bdev_set_options", 00:06:26.030 "accel_get_stats", 00:06:26.030 "accel_set_options", 00:06:26.030 "accel_set_driver", 00:06:26.030 "accel_crypto_key_destroy", 00:06:26.030 "accel_crypto_keys_get", 00:06:26.030 "accel_crypto_key_create", 00:06:26.030 "accel_assign_opc", 00:06:26.030 "accel_get_module_info", 00:06:26.030 "accel_get_opc_assignments", 00:06:26.030 "vmd_rescan", 00:06:26.030 "vmd_remove_device", 00:06:26.030 "vmd_enable", 00:06:26.030 "sock_get_default_impl", 00:06:26.030 "sock_set_default_impl", 00:06:26.030 "sock_impl_set_options", 00:06:26.030 "sock_impl_get_options", 00:06:26.030 "iobuf_get_stats", 00:06:26.030 "iobuf_set_options", 00:06:26.030 "keyring_get_keys", 00:06:26.030 "framework_get_pci_devices", 00:06:26.030 "framework_get_config", 00:06:26.030 "framework_get_subsystems", 00:06:26.030 "fsdev_set_opts", 00:06:26.030 "fsdev_get_opts", 00:06:26.030 "trace_get_info", 00:06:26.030 "trace_get_tpoint_group_mask", 00:06:26.030 "trace_disable_tpoint_group", 00:06:26.030 "trace_enable_tpoint_group", 00:06:26.030 "trace_clear_tpoint_mask", 00:06:26.030 "trace_set_tpoint_mask", 00:06:26.030 "notify_get_notifications", 00:06:26.030 "notify_get_types", 00:06:26.030 "spdk_get_version", 00:06:26.030 "rpc_get_methods" 00:06:26.030 ] 00:06:26.030 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:26.030 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:06:26.030 10:55:31 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 70059 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@952 -- # '[' -z 70059 ']' 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@956 -- # kill -0 70059 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@957 -- # uname 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70059 00:06:26.030 killing process with pid 70059 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70059' 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@971 -- # kill 70059 00:06:26.030 10:55:31 spdkcli_tcp -- common/autotest_common.sh@976 -- # wait 70059 00:06:26.598 00:06:26.598 real 0m1.817s 00:06:26.598 user 0m3.058s 00:06:26.598 sys 0m0.566s 00:06:26.598 10:55:31 spdkcli_tcp -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:26.598 10:55:31 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:26.598 ************************************ 00:06:26.598 END TEST spdkcli_tcp 00:06:26.598 ************************************ 00:06:26.598 10:55:31 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:26.598 10:55:31 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:26.598 10:55:31 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:26.598 10:55:31 -- common/autotest_common.sh@10 -- # set +x 00:06:26.598 ************************************ 00:06:26.598 START TEST dpdk_mem_utility 00:06:26.598 ************************************ 00:06:26.598 10:55:31 dpdk_mem_utility -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:26.598 * Looking for test storage... 00:06:26.598 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:06:26.598 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:26.598 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # lcov --version 00:06:26.598 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:26.858 10:55:32 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:26.858 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:26.858 --rc genhtml_branch_coverage=1 00:06:26.858 --rc genhtml_function_coverage=1 00:06:26.858 --rc genhtml_legend=1 00:06:26.858 --rc geninfo_all_blocks=1 00:06:26.858 --rc geninfo_unexecuted_blocks=1 00:06:26.858 00:06:26.858 ' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:26.858 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:26.858 --rc genhtml_branch_coverage=1 00:06:26.858 --rc genhtml_function_coverage=1 00:06:26.858 --rc genhtml_legend=1 00:06:26.858 --rc geninfo_all_blocks=1 00:06:26.858 --rc geninfo_unexecuted_blocks=1 00:06:26.858 00:06:26.858 ' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:26.858 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:26.858 --rc genhtml_branch_coverage=1 00:06:26.858 --rc genhtml_function_coverage=1 00:06:26.858 --rc genhtml_legend=1 00:06:26.858 --rc geninfo_all_blocks=1 00:06:26.858 --rc geninfo_unexecuted_blocks=1 00:06:26.858 00:06:26.858 ' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:26.858 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:26.858 --rc genhtml_branch_coverage=1 00:06:26.858 --rc genhtml_function_coverage=1 00:06:26.858 --rc genhtml_legend=1 00:06:26.858 --rc geninfo_all_blocks=1 00:06:26.858 --rc geninfo_unexecuted_blocks=1 00:06:26.858 00:06:26.858 ' 00:06:26.858 10:55:32 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:26.858 10:55:32 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=70148 00:06:26.858 10:55:32 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:26.858 10:55:32 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 70148 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@833 -- # '[' -z 70148 ']' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:26.858 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:26.858 10:55:32 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:26.858 [2024-10-29 10:55:32.274729] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:26.858 [2024-10-29 10:55:32.274917] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70148 ] 00:06:27.119 [2024-10-29 10:55:32.449139] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:27.119 [2024-10-29 10:55:32.474707] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:27.689 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:27.689 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@866 -- # return 0 00:06:27.689 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:06:27.689 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:06:27.689 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:27.689 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:27.689 { 00:06:27.689 "filename": "/tmp/spdk_mem_dump.txt" 00:06:27.689 } 00:06:27.689 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:27.689 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:27.689 DPDK memory size 810.000000 MiB in 1 heap(s) 00:06:27.689 1 heaps totaling size 810.000000 MiB 00:06:27.689 size: 810.000000 MiB heap id: 0 00:06:27.689 end heaps---------- 00:06:27.689 9 mempools totaling size 595.772034 MiB 00:06:27.689 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:06:27.689 size: 158.602051 MiB name: PDU_data_out_Pool 00:06:27.689 size: 92.545471 MiB name: bdev_io_70148 00:06:27.689 size: 50.003479 MiB name: msgpool_70148 00:06:27.689 size: 36.509338 MiB name: fsdev_io_70148 00:06:27.689 size: 21.763794 MiB name: PDU_Pool 00:06:27.689 size: 19.513306 MiB name: SCSI_TASK_Pool 00:06:27.689 size: 4.133484 MiB name: evtpool_70148 00:06:27.689 size: 0.026123 MiB name: Session_Pool 00:06:27.689 end mempools------- 00:06:27.689 6 memzones totaling size 4.142822 MiB 00:06:27.689 size: 1.000366 MiB name: RG_ring_0_70148 00:06:27.689 size: 1.000366 MiB name: RG_ring_1_70148 00:06:27.689 size: 1.000366 MiB name: RG_ring_4_70148 00:06:27.689 size: 1.000366 MiB name: RG_ring_5_70148 00:06:27.689 size: 0.125366 MiB name: RG_ring_2_70148 00:06:27.689 size: 0.015991 MiB name: RG_ring_3_70148 00:06:27.689 end memzones------- 00:06:27.689 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:06:27.951 heap id: 0 total size: 810.000000 MiB number of busy elements: 295 number of free elements: 15 00:06:27.951 list of free elements. size: 10.816528 MiB 00:06:27.951 element at address: 0x200018a00000 with size: 0.999878 MiB 00:06:27.951 element at address: 0x200018c00000 with size: 0.999878 MiB 00:06:27.951 element at address: 0x200031800000 with size: 0.994446 MiB 00:06:27.951 element at address: 0x200000400000 with size: 0.993958 MiB 00:06:27.951 element at address: 0x200006400000 with size: 0.959839 MiB 00:06:27.951 element at address: 0x200012c00000 with size: 0.954285 MiB 00:06:27.951 element at address: 0x200018e00000 with size: 0.936584 MiB 00:06:27.951 element at address: 0x200000200000 with size: 0.717346 MiB 00:06:27.951 element at address: 0x20001a600000 with size: 0.570435 MiB 00:06:27.951 element at address: 0x20000a600000 with size: 0.488892 MiB 00:06:27.951 element at address: 0x200000c00000 with size: 0.487000 MiB 00:06:27.951 element at address: 0x200019000000 with size: 0.485657 MiB 00:06:27.951 element at address: 0x200003e00000 with size: 0.480286 MiB 00:06:27.951 element at address: 0x200027a00000 with size: 0.396301 MiB 00:06:27.951 element at address: 0x200000800000 with size: 0.351746 MiB 00:06:27.951 list of standard malloc elements. size: 199.264587 MiB 00:06:27.951 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:06:27.951 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:06:27.951 element at address: 0x200018afff80 with size: 1.000122 MiB 00:06:27.951 element at address: 0x200018cfff80 with size: 1.000122 MiB 00:06:27.951 element at address: 0x200018efff80 with size: 1.000122 MiB 00:06:27.951 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:06:27.951 element at address: 0x200018eeff00 with size: 0.062622 MiB 00:06:27.951 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:06:27.951 element at address: 0x200018eefdc0 with size: 0.000305 MiB 00:06:27.951 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000085e580 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087e840 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087e900 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f080 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f140 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f200 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f380 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f440 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f500 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x20000087f680 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:06:27.951 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:06:27.951 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000cff000 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200003efb980 with size: 0.000183 MiB 00:06:27.952 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200012cf44c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200018eefc40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200018eefd00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x2000190bc740 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692080 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692140 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692200 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6922c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692380 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692440 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692500 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6925c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692680 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692740 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692800 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6928c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692980 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692a40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692b00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692bc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692c80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692d40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692e00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692ec0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a692f80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693040 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693100 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6931c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693280 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693340 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693400 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6934c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693580 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693640 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693700 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6937c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693880 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693940 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693a00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693ac0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693b80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693c40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693d00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693dc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693e80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a693f40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694000 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6940c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694180 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694240 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694300 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6943c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694480 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694540 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694600 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6946c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694780 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694840 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694900 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6949c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694a80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694b40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694c00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694cc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694d80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694e40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694f00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a694fc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a695080 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a695140 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a695200 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a6952c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a695380 with size: 0.000183 MiB 00:06:27.952 element at address: 0x20001a695440 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a65740 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a65800 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c400 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c600 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c6c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c780 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c840 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c900 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6c9c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6ca80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6cb40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6cc00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6ccc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6cd80 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6ce40 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6cf00 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6cfc0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d080 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d140 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d200 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d2c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d380 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d440 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d500 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d5c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d680 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d740 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d800 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d8c0 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6d980 with size: 0.000183 MiB 00:06:27.952 element at address: 0x200027a6da40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6db00 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6dbc0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6dc80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6dd40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6de00 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6dec0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6df80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e040 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e100 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e1c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e280 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e340 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e400 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e4c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e580 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e640 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e700 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e7c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e880 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6e940 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ea00 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6eac0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6eb80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ec40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ed00 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6edc0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ee80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ef40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f000 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f0c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f180 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f240 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f300 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f3c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f480 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f540 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f600 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f6c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f780 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f840 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f900 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6f9c0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fa80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fb40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fc00 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fcc0 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fd80 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6fe40 with size: 0.000183 MiB 00:06:27.953 element at address: 0x200027a6ff00 with size: 0.000183 MiB 00:06:27.953 list of memzone associated elements. size: 599.918884 MiB 00:06:27.953 element at address: 0x20001a695500 with size: 211.416748 MiB 00:06:27.953 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:06:27.953 element at address: 0x200027a6ffc0 with size: 157.562561 MiB 00:06:27.953 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:06:27.953 element at address: 0x200012df4780 with size: 92.045044 MiB 00:06:27.953 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_70148_0 00:06:27.953 element at address: 0x200000dff380 with size: 48.003052 MiB 00:06:27.953 associated memzone info: size: 48.002930 MiB name: MP_msgpool_70148_0 00:06:27.953 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:06:27.953 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_70148_0 00:06:27.953 element at address: 0x2000191be940 with size: 20.255554 MiB 00:06:27.953 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:06:27.953 element at address: 0x2000319feb40 with size: 18.005066 MiB 00:06:27.953 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:06:27.953 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:06:27.953 associated memzone info: size: 3.000122 MiB name: MP_evtpool_70148_0 00:06:27.953 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:06:27.953 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_70148 00:06:27.953 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:06:27.953 associated memzone info: size: 1.007996 MiB name: MP_evtpool_70148 00:06:27.953 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:06:27.953 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:06:27.953 element at address: 0x2000190bc800 with size: 1.008118 MiB 00:06:27.953 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:06:27.953 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:06:27.953 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:06:27.953 element at address: 0x200003efba40 with size: 1.008118 MiB 00:06:27.953 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:06:27.953 element at address: 0x200000cff180 with size: 1.000488 MiB 00:06:27.953 associated memzone info: size: 1.000366 MiB name: RG_ring_0_70148 00:06:27.953 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:06:27.953 associated memzone info: size: 1.000366 MiB name: RG_ring_1_70148 00:06:27.953 element at address: 0x200012cf4580 with size: 1.000488 MiB 00:06:27.953 associated memzone info: size: 1.000366 MiB name: RG_ring_4_70148 00:06:27.953 element at address: 0x2000318fe940 with size: 1.000488 MiB 00:06:27.953 associated memzone info: size: 1.000366 MiB name: RG_ring_5_70148 00:06:27.953 element at address: 0x20000087f740 with size: 0.500488 MiB 00:06:27.953 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_70148 00:06:27.953 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:06:27.953 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_70148 00:06:27.953 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:06:27.953 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:06:27.953 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:06:27.953 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:06:27.953 element at address: 0x20001907c540 with size: 0.250488 MiB 00:06:27.953 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:06:27.953 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:06:27.953 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_70148 00:06:27.953 element at address: 0x20000085e640 with size: 0.125488 MiB 00:06:27.953 associated memzone info: size: 0.125366 MiB name: RG_ring_2_70148 00:06:27.953 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:06:27.953 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:06:27.953 element at address: 0x200027a658c0 with size: 0.023743 MiB 00:06:27.953 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:06:27.953 element at address: 0x20000085a380 with size: 0.016113 MiB 00:06:27.953 associated memzone info: size: 0.015991 MiB name: RG_ring_3_70148 00:06:27.953 element at address: 0x200027a6ba00 with size: 0.002441 MiB 00:06:27.953 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:06:27.953 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:06:27.953 associated memzone info: size: 0.000183 MiB name: MP_msgpool_70148 00:06:27.953 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:06:27.953 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_70148 00:06:27.953 element at address: 0x20000085a180 with size: 0.000305 MiB 00:06:27.953 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_70148 00:06:27.953 element at address: 0x200027a6c4c0 with size: 0.000305 MiB 00:06:27.953 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:06:27.953 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:06:27.953 10:55:33 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 70148 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@952 -- # '[' -z 70148 ']' 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@956 -- # kill -0 70148 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@957 -- # uname 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70148 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70148' 00:06:27.953 killing process with pid 70148 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@971 -- # kill 70148 00:06:27.953 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@976 -- # wait 70148 00:06:28.213 00:06:28.213 real 0m1.654s 00:06:28.213 user 0m1.580s 00:06:28.213 sys 0m0.529s 00:06:28.213 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:28.213 10:55:33 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:28.213 ************************************ 00:06:28.213 END TEST dpdk_mem_utility 00:06:28.213 ************************************ 00:06:28.213 10:55:33 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:28.213 10:55:33 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:28.213 10:55:33 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:28.213 10:55:33 -- common/autotest_common.sh@10 -- # set +x 00:06:28.213 ************************************ 00:06:28.213 START TEST event 00:06:28.213 ************************************ 00:06:28.213 10:55:33 event -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:28.474 * Looking for test storage... 00:06:28.474 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1691 -- # lcov --version 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:28.474 10:55:33 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:28.474 10:55:33 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:28.474 10:55:33 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:28.474 10:55:33 event -- scripts/common.sh@336 -- # IFS=.-: 00:06:28.474 10:55:33 event -- scripts/common.sh@336 -- # read -ra ver1 00:06:28.474 10:55:33 event -- scripts/common.sh@337 -- # IFS=.-: 00:06:28.474 10:55:33 event -- scripts/common.sh@337 -- # read -ra ver2 00:06:28.474 10:55:33 event -- scripts/common.sh@338 -- # local 'op=<' 00:06:28.474 10:55:33 event -- scripts/common.sh@340 -- # ver1_l=2 00:06:28.474 10:55:33 event -- scripts/common.sh@341 -- # ver2_l=1 00:06:28.474 10:55:33 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:28.474 10:55:33 event -- scripts/common.sh@344 -- # case "$op" in 00:06:28.474 10:55:33 event -- scripts/common.sh@345 -- # : 1 00:06:28.474 10:55:33 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:28.474 10:55:33 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:28.474 10:55:33 event -- scripts/common.sh@365 -- # decimal 1 00:06:28.474 10:55:33 event -- scripts/common.sh@353 -- # local d=1 00:06:28.474 10:55:33 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:28.474 10:55:33 event -- scripts/common.sh@355 -- # echo 1 00:06:28.474 10:55:33 event -- scripts/common.sh@365 -- # ver1[v]=1 00:06:28.474 10:55:33 event -- scripts/common.sh@366 -- # decimal 2 00:06:28.474 10:55:33 event -- scripts/common.sh@353 -- # local d=2 00:06:28.474 10:55:33 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:28.474 10:55:33 event -- scripts/common.sh@355 -- # echo 2 00:06:28.474 10:55:33 event -- scripts/common.sh@366 -- # ver2[v]=2 00:06:28.474 10:55:33 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:28.474 10:55:33 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:28.474 10:55:33 event -- scripts/common.sh@368 -- # return 0 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:28.474 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:28.474 --rc genhtml_branch_coverage=1 00:06:28.474 --rc genhtml_function_coverage=1 00:06:28.474 --rc genhtml_legend=1 00:06:28.474 --rc geninfo_all_blocks=1 00:06:28.474 --rc geninfo_unexecuted_blocks=1 00:06:28.474 00:06:28.474 ' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:28.474 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:28.474 --rc genhtml_branch_coverage=1 00:06:28.474 --rc genhtml_function_coverage=1 00:06:28.474 --rc genhtml_legend=1 00:06:28.474 --rc geninfo_all_blocks=1 00:06:28.474 --rc geninfo_unexecuted_blocks=1 00:06:28.474 00:06:28.474 ' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:28.474 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:28.474 --rc genhtml_branch_coverage=1 00:06:28.474 --rc genhtml_function_coverage=1 00:06:28.474 --rc genhtml_legend=1 00:06:28.474 --rc geninfo_all_blocks=1 00:06:28.474 --rc geninfo_unexecuted_blocks=1 00:06:28.474 00:06:28.474 ' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:28.474 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:28.474 --rc genhtml_branch_coverage=1 00:06:28.474 --rc genhtml_function_coverage=1 00:06:28.474 --rc genhtml_legend=1 00:06:28.474 --rc geninfo_all_blocks=1 00:06:28.474 --rc geninfo_unexecuted_blocks=1 00:06:28.474 00:06:28.474 ' 00:06:28.474 10:55:33 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:28.474 10:55:33 event -- bdev/nbd_common.sh@6 -- # set -e 00:06:28.474 10:55:33 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1103 -- # '[' 6 -le 1 ']' 00:06:28.474 10:55:33 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:28.474 10:55:33 event -- common/autotest_common.sh@10 -- # set +x 00:06:28.474 ************************************ 00:06:28.474 START TEST event_perf 00:06:28.474 ************************************ 00:06:28.474 10:55:33 event.event_perf -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:28.474 Running I/O for 1 seconds...[2024-10-29 10:55:33.956527] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:28.474 [2024-10-29 10:55:33.956672] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70234 ] 00:06:28.734 [2024-10-29 10:55:34.127602] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:28.734 [2024-10-29 10:55:34.156751] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:28.734 [2024-10-29 10:55:34.156878] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:28.734 Running I/O for 1 seconds...[2024-10-29 10:55:34.157094] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:28.734 [2024-10-29 10:55:34.156968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:30.112 00:06:30.112 lcore 0: 96439 00:06:30.112 lcore 1: 96441 00:06:30.112 lcore 2: 96444 00:06:30.112 lcore 3: 96441 00:06:30.112 done. 00:06:30.112 00:06:30.112 real 0m1.337s 00:06:30.112 user 0m4.116s 00:06:30.112 sys 0m0.099s 00:06:30.112 10:55:35 event.event_perf -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:30.112 10:55:35 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:06:30.112 ************************************ 00:06:30.112 END TEST event_perf 00:06:30.112 ************************************ 00:06:30.112 10:55:35 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:30.112 10:55:35 event -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:06:30.112 10:55:35 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:30.112 10:55:35 event -- common/autotest_common.sh@10 -- # set +x 00:06:30.112 ************************************ 00:06:30.112 START TEST event_reactor 00:06:30.112 ************************************ 00:06:30.112 10:55:35 event.event_reactor -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:30.112 [2024-10-29 10:55:35.360170] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:30.112 [2024-10-29 10:55:35.360305] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70268 ] 00:06:30.112 [2024-10-29 10:55:35.528827] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:30.112 [2024-10-29 10:55:35.569400] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:31.487 test_start 00:06:31.487 oneshot 00:06:31.487 tick 100 00:06:31.487 tick 100 00:06:31.487 tick 250 00:06:31.487 tick 100 00:06:31.487 tick 100 00:06:31.487 tick 250 00:06:31.487 tick 100 00:06:31.487 tick 500 00:06:31.487 tick 100 00:06:31.487 tick 100 00:06:31.487 tick 250 00:06:31.487 tick 100 00:06:31.487 tick 100 00:06:31.487 test_end 00:06:31.487 00:06:31.487 real 0m1.337s 00:06:31.487 user 0m1.129s 00:06:31.487 sys 0m0.098s 00:06:31.487 10:55:36 event.event_reactor -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:31.487 10:55:36 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:06:31.487 ************************************ 00:06:31.487 END TEST event_reactor 00:06:31.487 ************************************ 00:06:31.487 10:55:36 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:31.487 10:55:36 event -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:06:31.487 10:55:36 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:31.487 10:55:36 event -- common/autotest_common.sh@10 -- # set +x 00:06:31.487 ************************************ 00:06:31.487 START TEST event_reactor_perf 00:06:31.487 ************************************ 00:06:31.487 10:55:36 event.event_reactor_perf -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:31.487 [2024-10-29 10:55:36.763417] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:31.488 [2024-10-29 10:55:36.763535] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70310 ] 00:06:31.488 [2024-10-29 10:55:36.931878] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:31.488 [2024-10-29 10:55:36.969436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:32.867 test_start 00:06:32.867 test_end 00:06:32.867 Performance: 399876 events per second 00:06:32.867 00:06:32.867 real 0m1.335s 00:06:32.867 user 0m1.129s 00:06:32.867 sys 0m0.097s 00:06:32.867 10:55:38 event.event_reactor_perf -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:32.868 10:55:38 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:06:32.868 ************************************ 00:06:32.868 END TEST event_reactor_perf 00:06:32.868 ************************************ 00:06:32.868 10:55:38 event -- event/event.sh@49 -- # uname -s 00:06:32.868 10:55:38 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:06:32.868 10:55:38 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:32.868 10:55:38 event -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:32.868 10:55:38 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:32.868 10:55:38 event -- common/autotest_common.sh@10 -- # set +x 00:06:32.868 ************************************ 00:06:32.868 START TEST event_scheduler 00:06:32.868 ************************************ 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:32.868 * Looking for test storage... 00:06:32.868 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1691 -- # lcov --version 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:32.868 10:55:38 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:32.868 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:32.868 --rc genhtml_branch_coverage=1 00:06:32.868 --rc genhtml_function_coverage=1 00:06:32.868 --rc genhtml_legend=1 00:06:32.868 --rc geninfo_all_blocks=1 00:06:32.868 --rc geninfo_unexecuted_blocks=1 00:06:32.868 00:06:32.868 ' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:32.868 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:32.868 --rc genhtml_branch_coverage=1 00:06:32.868 --rc genhtml_function_coverage=1 00:06:32.868 --rc genhtml_legend=1 00:06:32.868 --rc geninfo_all_blocks=1 00:06:32.868 --rc geninfo_unexecuted_blocks=1 00:06:32.868 00:06:32.868 ' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:32.868 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:32.868 --rc genhtml_branch_coverage=1 00:06:32.868 --rc genhtml_function_coverage=1 00:06:32.868 --rc genhtml_legend=1 00:06:32.868 --rc geninfo_all_blocks=1 00:06:32.868 --rc geninfo_unexecuted_blocks=1 00:06:32.868 00:06:32.868 ' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:32.868 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:32.868 --rc genhtml_branch_coverage=1 00:06:32.868 --rc genhtml_function_coverage=1 00:06:32.868 --rc genhtml_legend=1 00:06:32.868 --rc geninfo_all_blocks=1 00:06:32.868 --rc geninfo_unexecuted_blocks=1 00:06:32.868 00:06:32.868 ' 00:06:32.868 10:55:38 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:06:32.868 10:55:38 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=70375 00:06:32.868 10:55:38 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:06:32.868 10:55:38 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:06:32.868 10:55:38 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 70375 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@833 -- # '[' -z 70375 ']' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:32.868 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:32.868 10:55:38 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:33.128 [2024-10-29 10:55:38.432158] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:33.128 [2024-10-29 10:55:38.432297] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70375 ] 00:06:33.128 [2024-10-29 10:55:38.623008] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:33.389 [2024-10-29 10:55:38.653302] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:33.389 [2024-10-29 10:55:38.655876] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:33.389 [2024-10-29 10:55:38.655962] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:33.389 [2024-10-29 10:55:38.656580] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@866 -- # return 0 00:06:33.957 10:55:39 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:33.957 POWER: Cannot set governor of lcore 0 to userspace 00:06:33.957 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:33.957 POWER: Cannot set governor of lcore 0 to performance 00:06:33.957 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:33.957 POWER: Cannot set governor of lcore 0 to userspace 00:06:33.957 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:33.957 POWER: Cannot set governor of lcore 0 to userspace 00:06:33.957 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:06:33.957 POWER: Unable to set Power Management Environment for lcore 0 00:06:33.957 [2024-10-29 10:55:39.260695] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:06:33.957 [2024-10-29 10:55:39.260718] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:06:33.957 [2024-10-29 10:55:39.260729] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:06:33.957 [2024-10-29 10:55:39.260746] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:06:33.957 [2024-10-29 10:55:39.260767] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:06:33.957 [2024-10-29 10:55:39.260777] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.957 10:55:39 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 [2024-10-29 10:55:39.330972] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.957 10:55:39 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 ************************************ 00:06:33.957 START TEST scheduler_create_thread 00:06:33.957 ************************************ 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1127 -- # scheduler_create_thread 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 2 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 3 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.957 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.957 4 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 5 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 6 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 7 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 8 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 9 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 10 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:33.958 10:55:39 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:35.432 10:55:40 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:35.432 10:55:40 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:35.432 10:55:40 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:35.432 10:55:40 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:35.432 10:55:40 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:36.805 10:55:41 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:36.805 00:06:36.805 real 0m2.609s 00:06:36.805 user 0m0.007s 00:06:36.805 sys 0m0.011s 00:06:36.805 10:55:41 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:36.805 10:55:41 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:36.805 ************************************ 00:06:36.805 END TEST scheduler_create_thread 00:06:36.805 ************************************ 00:06:36.805 10:55:41 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:36.805 10:55:41 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 70375 00:06:36.805 10:55:41 event.event_scheduler -- common/autotest_common.sh@952 -- # '[' -z 70375 ']' 00:06:36.805 10:55:41 event.event_scheduler -- common/autotest_common.sh@956 -- # kill -0 70375 00:06:36.805 10:55:41 event.event_scheduler -- common/autotest_common.sh@957 -- # uname 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70375 00:06:36.805 killing process with pid 70375 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@958 -- # process_name=reactor_2 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@962 -- # '[' reactor_2 = sudo ']' 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70375' 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@971 -- # kill 70375 00:06:36.805 10:55:42 event.event_scheduler -- common/autotest_common.sh@976 -- # wait 70375 00:06:37.063 [2024-10-29 10:55:42.429503] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:37.321 00:06:37.321 real 0m4.520s 00:06:37.321 user 0m7.961s 00:06:37.321 sys 0m0.480s 00:06:37.321 10:55:42 event.event_scheduler -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:37.321 10:55:42 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:37.321 ************************************ 00:06:37.321 END TEST event_scheduler 00:06:37.321 ************************************ 00:06:37.321 10:55:42 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:37.321 10:55:42 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:37.321 10:55:42 event -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:37.321 10:55:42 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:37.321 10:55:42 event -- common/autotest_common.sh@10 -- # set +x 00:06:37.321 ************************************ 00:06:37.321 START TEST app_repeat 00:06:37.321 ************************************ 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@1127 -- # app_repeat_test 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70476 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70476' 00:06:37.321 Process app_repeat pid: 70476 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:37.321 spdk_app_start Round 0 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:37.321 10:55:42 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70476 /var/tmp/spdk-nbd.sock 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@833 -- # '[' -z 70476 ']' 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:37.321 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:37.321 10:55:42 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:37.321 [2024-10-29 10:55:42.778256] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:37.321 [2024-10-29 10:55:42.778373] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70476 ] 00:06:37.579 [2024-10-29 10:55:42.949188] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:37.579 [2024-10-29 10:55:42.978367] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:37.579 [2024-10-29 10:55:42.978447] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:38.143 10:55:43 event.app_repeat -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:38.143 10:55:43 event.app_repeat -- common/autotest_common.sh@866 -- # return 0 00:06:38.143 10:55:43 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:38.401 Malloc0 00:06:38.401 10:55:43 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:38.659 Malloc1 00:06:38.660 10:55:44 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:38.660 10:55:44 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:38.918 /dev/nbd0 00:06:38.918 10:55:44 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:38.918 10:55:44 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:38.918 1+0 records in 00:06:38.918 1+0 records out 00:06:38.918 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000351933 s, 11.6 MB/s 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:38.918 10:55:44 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:38.918 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:38.918 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:38.918 10:55:44 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:39.176 /dev/nbd1 00:06:39.176 10:55:44 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:39.176 10:55:44 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:39.176 10:55:44 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:06:39.176 10:55:44 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:39.176 10:55:44 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:39.176 10:55:44 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:39.177 1+0 records in 00:06:39.177 1+0 records out 00:06:39.177 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000353717 s, 11.6 MB/s 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:39.177 10:55:44 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:39.177 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:39.177 10:55:44 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:39.177 10:55:44 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:39.177 10:55:44 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:39.177 10:55:44 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:39.434 10:55:44 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:39.434 { 00:06:39.434 "nbd_device": "/dev/nbd0", 00:06:39.434 "bdev_name": "Malloc0" 00:06:39.434 }, 00:06:39.434 { 00:06:39.434 "nbd_device": "/dev/nbd1", 00:06:39.434 "bdev_name": "Malloc1" 00:06:39.434 } 00:06:39.434 ]' 00:06:39.434 10:55:44 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:39.434 { 00:06:39.434 "nbd_device": "/dev/nbd0", 00:06:39.434 "bdev_name": "Malloc0" 00:06:39.435 }, 00:06:39.435 { 00:06:39.435 "nbd_device": "/dev/nbd1", 00:06:39.435 "bdev_name": "Malloc1" 00:06:39.435 } 00:06:39.435 ]' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:39.435 /dev/nbd1' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:39.435 /dev/nbd1' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:39.435 256+0 records in 00:06:39.435 256+0 records out 00:06:39.435 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0129828 s, 80.8 MB/s 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:39.435 256+0 records in 00:06:39.435 256+0 records out 00:06:39.435 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0174157 s, 60.2 MB/s 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:39.435 256+0 records in 00:06:39.435 256+0 records out 00:06:39.435 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.016522 s, 63.5 MB/s 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:39.435 10:55:44 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:39.693 10:55:45 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:39.951 10:55:45 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:40.208 10:55:45 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:40.208 10:55:45 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:40.465 10:55:45 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:40.465 [2024-10-29 10:55:45.909484] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:40.465 [2024-10-29 10:55:45.933340] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:40.465 [2024-10-29 10:55:45.933342] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:40.723 [2024-10-29 10:55:45.976774] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:40.723 [2024-10-29 10:55:45.976912] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:44.008 10:55:48 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:44.008 10:55:48 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:44.008 spdk_app_start Round 1 00:06:44.008 10:55:48 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70476 /var/tmp/spdk-nbd.sock 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@833 -- # '[' -z 70476 ']' 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:44.008 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:44.008 10:55:48 event.app_repeat -- common/autotest_common.sh@866 -- # return 0 00:06:44.008 10:55:48 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:44.008 Malloc0 00:06:44.008 10:55:49 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:44.008 Malloc1 00:06:44.008 10:55:49 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:44.008 10:55:49 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:44.267 /dev/nbd0 00:06:44.267 10:55:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:44.267 10:55:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:44.267 1+0 records in 00:06:44.267 1+0 records out 00:06:44.267 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00024371 s, 16.8 MB/s 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:44.267 10:55:49 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:44.267 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:44.267 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:44.267 10:55:49 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:44.525 /dev/nbd1 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:44.525 1+0 records in 00:06:44.525 1+0 records out 00:06:44.525 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000229352 s, 17.9 MB/s 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:44.525 10:55:49 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:44.525 10:55:49 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:44.783 { 00:06:44.783 "nbd_device": "/dev/nbd0", 00:06:44.783 "bdev_name": "Malloc0" 00:06:44.783 }, 00:06:44.783 { 00:06:44.783 "nbd_device": "/dev/nbd1", 00:06:44.783 "bdev_name": "Malloc1" 00:06:44.783 } 00:06:44.783 ]' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:44.783 { 00:06:44.783 "nbd_device": "/dev/nbd0", 00:06:44.783 "bdev_name": "Malloc0" 00:06:44.783 }, 00:06:44.783 { 00:06:44.783 "nbd_device": "/dev/nbd1", 00:06:44.783 "bdev_name": "Malloc1" 00:06:44.783 } 00:06:44.783 ]' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:44.783 /dev/nbd1' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:44.783 /dev/nbd1' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:44.783 10:55:50 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:44.784 256+0 records in 00:06:44.784 256+0 records out 00:06:44.784 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.015344 s, 68.3 MB/s 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:44.784 256+0 records in 00:06:44.784 256+0 records out 00:06:44.784 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0187598 s, 55.9 MB/s 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:44.784 256+0 records in 00:06:44.784 256+0 records out 00:06:44.784 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0235339 s, 44.6 MB/s 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:44.784 10:55:50 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:45.042 10:55:50 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:45.301 10:55:50 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:45.559 10:55:50 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:45.559 10:55:50 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:45.818 10:55:51 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:45.818 [2024-10-29 10:55:51.263077] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:45.818 [2024-10-29 10:55:51.287031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:45.818 [2024-10-29 10:55:51.287058] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:46.076 [2024-10-29 10:55:51.329948] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:46.076 [2024-10-29 10:55:51.330017] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:49.362 spdk_app_start Round 2 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70476 /var/tmp/spdk-nbd.sock 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@833 -- # '[' -z 70476 ']' 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:49.362 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:49.362 10:55:54 event.app_repeat -- common/autotest_common.sh@866 -- # return 0 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:49.362 Malloc0 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:49.362 Malloc1 00:06:49.362 10:55:54 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:49.362 10:55:54 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:49.621 /dev/nbd0 00:06:49.621 10:55:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:49.621 10:55:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:49.621 1+0 records in 00:06:49.621 1+0 records out 00:06:49.621 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000309233 s, 13.2 MB/s 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:49.621 10:55:55 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:49.621 10:55:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:49.621 10:55:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:49.621 10:55:55 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:49.879 /dev/nbd1 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@871 -- # local i 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@875 -- # break 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:49.879 1+0 records in 00:06:49.879 1+0 records out 00:06:49.879 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000305101 s, 13.4 MB/s 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@888 -- # size=4096 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:06:49.879 10:55:55 event.app_repeat -- common/autotest_common.sh@891 -- # return 0 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:49.879 10:55:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:50.137 10:55:55 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:50.137 { 00:06:50.137 "nbd_device": "/dev/nbd0", 00:06:50.137 "bdev_name": "Malloc0" 00:06:50.137 }, 00:06:50.137 { 00:06:50.137 "nbd_device": "/dev/nbd1", 00:06:50.137 "bdev_name": "Malloc1" 00:06:50.137 } 00:06:50.137 ]' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:50.138 { 00:06:50.138 "nbd_device": "/dev/nbd0", 00:06:50.138 "bdev_name": "Malloc0" 00:06:50.138 }, 00:06:50.138 { 00:06:50.138 "nbd_device": "/dev/nbd1", 00:06:50.138 "bdev_name": "Malloc1" 00:06:50.138 } 00:06:50.138 ]' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:50.138 /dev/nbd1' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:50.138 /dev/nbd1' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:50.138 256+0 records in 00:06:50.138 256+0 records out 00:06:50.138 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0141866 s, 73.9 MB/s 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:50.138 256+0 records in 00:06:50.138 256+0 records out 00:06:50.138 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0212389 s, 49.4 MB/s 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:50.138 256+0 records in 00:06:50.138 256+0 records out 00:06:50.138 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0218362 s, 48.0 MB/s 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:50.138 10:55:55 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:50.396 10:55:55 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:50.654 10:55:56 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:50.912 10:55:56 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:50.912 10:55:56 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:51.170 10:55:56 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:51.170 [2024-10-29 10:55:56.622567] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:51.170 [2024-10-29 10:55:56.647075] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:51.170 [2024-10-29 10:55:56.647076] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:51.428 [2024-10-29 10:55:56.689815] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:51.428 [2024-10-29 10:55:56.689877] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:54.743 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:54.744 10:55:59 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70476 /var/tmp/spdk-nbd.sock 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@833 -- # '[' -z 70476 ']' 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@866 -- # return 0 00:06:54.744 10:55:59 event.app_repeat -- event/event.sh@39 -- # killprocess 70476 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@952 -- # '[' -z 70476 ']' 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@956 -- # kill -0 70476 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@957 -- # uname 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70476 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70476' 00:06:54.744 killing process with pid 70476 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@971 -- # kill 70476 00:06:54.744 10:55:59 event.app_repeat -- common/autotest_common.sh@976 -- # wait 70476 00:06:54.744 spdk_app_start is called in Round 0. 00:06:54.744 Shutdown signal received, stop current app iteration 00:06:54.744 Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 reinitialization... 00:06:54.744 spdk_app_start is called in Round 1. 00:06:54.744 Shutdown signal received, stop current app iteration 00:06:54.744 Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 reinitialization... 00:06:54.744 spdk_app_start is called in Round 2. 00:06:54.744 Shutdown signal received, stop current app iteration 00:06:54.744 Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 reinitialization... 00:06:54.744 spdk_app_start is called in Round 3. 00:06:54.744 Shutdown signal received, stop current app iteration 00:06:54.744 10:56:00 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:54.744 10:56:00 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:54.744 00:06:54.744 real 0m17.306s 00:06:54.744 user 0m38.154s 00:06:54.744 sys 0m2.617s 00:06:54.744 10:56:00 event.app_repeat -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:54.744 10:56:00 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:54.744 ************************************ 00:06:54.744 END TEST app_repeat 00:06:54.744 ************************************ 00:06:54.744 10:56:00 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:54.744 10:56:00 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:54.744 10:56:00 event -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:54.744 10:56:00 event -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:54.744 10:56:00 event -- common/autotest_common.sh@10 -- # set +x 00:06:54.744 ************************************ 00:06:54.744 START TEST cpu_locks 00:06:54.744 ************************************ 00:06:54.744 10:56:00 event.cpu_locks -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:54.744 * Looking for test storage... 00:06:54.744 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:54.744 10:56:00 event.cpu_locks -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:06:54.744 10:56:00 event.cpu_locks -- common/autotest_common.sh@1691 -- # lcov --version 00:06:54.744 10:56:00 event.cpu_locks -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:55.005 10:56:00 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:06:55.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:55.005 --rc genhtml_branch_coverage=1 00:06:55.005 --rc genhtml_function_coverage=1 00:06:55.005 --rc genhtml_legend=1 00:06:55.005 --rc geninfo_all_blocks=1 00:06:55.005 --rc geninfo_unexecuted_blocks=1 00:06:55.005 00:06:55.005 ' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:06:55.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:55.005 --rc genhtml_branch_coverage=1 00:06:55.005 --rc genhtml_function_coverage=1 00:06:55.005 --rc genhtml_legend=1 00:06:55.005 --rc geninfo_all_blocks=1 00:06:55.005 --rc geninfo_unexecuted_blocks=1 00:06:55.005 00:06:55.005 ' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:06:55.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:55.005 --rc genhtml_branch_coverage=1 00:06:55.005 --rc genhtml_function_coverage=1 00:06:55.005 --rc genhtml_legend=1 00:06:55.005 --rc geninfo_all_blocks=1 00:06:55.005 --rc geninfo_unexecuted_blocks=1 00:06:55.005 00:06:55.005 ' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:06:55.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:55.005 --rc genhtml_branch_coverage=1 00:06:55.005 --rc genhtml_function_coverage=1 00:06:55.005 --rc genhtml_legend=1 00:06:55.005 --rc geninfo_all_blocks=1 00:06:55.005 --rc geninfo_unexecuted_blocks=1 00:06:55.005 00:06:55.005 ' 00:06:55.005 10:56:00 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:55.005 10:56:00 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:55.005 10:56:00 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:55.005 10:56:00 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:55.005 10:56:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:55.005 ************************************ 00:06:55.006 START TEST default_locks 00:06:55.006 ************************************ 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@1127 -- # default_locks 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70903 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70903 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@833 -- # '[' -z 70903 ']' 00:06:55.006 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:55.006 10:56:00 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:55.006 [2024-10-29 10:56:00.440658] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:55.006 [2024-10-29 10:56:00.440815] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70903 ] 00:06:55.267 [2024-10-29 10:56:00.616882] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:55.267 [2024-10-29 10:56:00.657418] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:55.836 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:55.836 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@866 -- # return 0 00:06:55.836 10:56:01 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70903 00:06:55.836 10:56:01 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70903 00:06:55.836 10:56:01 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:56.405 10:56:01 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70903 00:06:56.405 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@952 -- # '[' -z 70903 ']' 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # kill -0 70903 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@957 -- # uname 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70903 00:06:56.406 killing process with pid 70903 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70903' 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@971 -- # kill 70903 00:06:56.406 10:56:01 event.cpu_locks.default_locks -- common/autotest_common.sh@976 -- # wait 70903 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70903 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70903 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 70903 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@833 -- # '[' -z 70903 ']' 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.976 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:56.976 ERROR: process (pid: 70903) is no longer running 00:06:56.976 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 848: kill: (70903) - No such process 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@866 -- # return 1 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:56.976 ************************************ 00:06:56.976 END TEST default_locks 00:06:56.976 ************************************ 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:56.976 00:06:56.976 real 0m1.982s 00:06:56.976 user 0m1.784s 00:06:56.976 sys 0m0.773s 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:56.976 10:56:02 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:56.976 10:56:02 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:56.976 10:56:02 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:56.976 10:56:02 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:56.976 10:56:02 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:56.976 ************************************ 00:06:56.976 START TEST default_locks_via_rpc 00:06:56.976 ************************************ 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1127 -- # default_locks_via_rpc 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70956 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70956 00:06:56.976 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@833 -- # '[' -z 70956 ']' 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:56.976 10:56:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:57.236 [2024-10-29 10:56:02.479961] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:57.236 [2024-10-29 10:56:02.480168] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70956 ] 00:06:57.236 [2024-10-29 10:56:02.653135] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:57.236 [2024-10-29 10:56:02.698536] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@866 -- # return 0 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.807 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:58.068 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.068 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70956 00:06:58.068 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70956 00:06:58.068 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70956 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@952 -- # '[' -z 70956 ']' 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # kill -0 70956 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@957 -- # uname 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 70956 00:06:58.329 killing process with pid 70956 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@970 -- # echo 'killing process with pid 70956' 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@971 -- # kill 70956 00:06:58.329 10:56:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@976 -- # wait 70956 00:06:58.899 00:06:58.899 real 0m1.895s 00:06:58.899 user 0m1.702s 00:06:58.899 sys 0m0.724s 00:06:58.899 10:56:04 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:06:58.899 ************************************ 00:06:58.899 END TEST default_locks_via_rpc 00:06:58.899 ************************************ 00:06:58.899 10:56:04 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:58.899 10:56:04 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:58.899 10:56:04 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:06:58.899 10:56:04 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:06:58.899 10:56:04 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:58.899 ************************************ 00:06:58.899 START TEST non_locking_app_on_locked_coremask 00:06:58.899 ************************************ 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1127 -- # non_locking_app_on_locked_coremask 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=71008 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 71008 /var/tmp/spdk.sock 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71008 ']' 00:06:58.899 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:58.899 10:56:04 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:59.160 [2024-10-29 10:56:04.443487] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:59.160 [2024-10-29 10:56:04.443616] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71008 ] 00:06:59.160 [2024-10-29 10:56:04.617625] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:59.420 [2024-10-29 10:56:04.662995] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@866 -- # return 0 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=71019 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 71019 /var/tmp/spdk2.sock 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71019 ']' 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:59.991 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:06:59.991 10:56:05 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:59.991 [2024-10-29 10:56:05.333452] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:06:59.991 [2024-10-29 10:56:05.333674] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71019 ] 00:07:00.251 [2024-10-29 10:56:05.504557] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:00.251 [2024-10-29 10:56:05.504660] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:00.251 [2024-10-29 10:56:05.593504] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:00.821 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:00.821 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@866 -- # return 0 00:07:00.821 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 71008 00:07:00.821 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71008 00:07:00.821 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 71008 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # '[' -z 71008 ']' 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # kill -0 71008 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # uname 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71008 00:07:01.391 killing process with pid 71008 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71008' 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@971 -- # kill 71008 00:07:01.391 10:56:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@976 -- # wait 71008 00:07:02.773 10:56:07 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 71019 00:07:02.773 10:56:07 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # '[' -z 71019 ']' 00:07:02.773 10:56:07 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # kill -0 71019 00:07:02.773 10:56:07 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # uname 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71019 00:07:02.773 killing process with pid 71019 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71019' 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@971 -- # kill 71019 00:07:02.773 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@976 -- # wait 71019 00:07:03.346 00:07:03.346 real 0m4.312s 00:07:03.346 user 0m4.162s 00:07:03.346 sys 0m1.363s 00:07:03.346 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:03.346 ************************************ 00:07:03.346 END TEST non_locking_app_on_locked_coremask 00:07:03.346 ************************************ 00:07:03.346 10:56:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:03.346 10:56:08 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:07:03.346 10:56:08 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:03.346 10:56:08 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:03.346 10:56:08 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:03.346 ************************************ 00:07:03.346 START TEST locking_app_on_unlocked_coremask 00:07:03.346 ************************************ 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1127 -- # locking_app_on_unlocked_coremask 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=71094 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 71094 /var/tmp/spdk.sock 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71094 ']' 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.346 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:03.346 10:56:08 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:03.346 [2024-10-29 10:56:08.823817] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:03.346 [2024-10-29 10:56:08.824064] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71094 ] 00:07:03.608 [2024-10-29 10:56:08.996514] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:03.608 [2024-10-29 10:56:08.996743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:03.608 [2024-10-29 10:56:09.040512] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@866 -- # return 0 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=71110 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 71110 /var/tmp/spdk2.sock 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71110 ']' 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:04.178 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:04.178 10:56:09 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:04.435 [2024-10-29 10:56:09.714079] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:04.435 [2024-10-29 10:56:09.714293] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71110 ] 00:07:04.435 [2024-10-29 10:56:09.880541] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:04.693 [2024-10-29 10:56:09.968859] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:05.260 10:56:10 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:05.260 10:56:10 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@866 -- # return 0 00:07:05.260 10:56:10 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 71110 00:07:05.260 10:56:10 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71110 00:07:05.260 10:56:10 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:05.520 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 71094 00:07:05.520 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@952 -- # '[' -z 71094 ']' 00:07:05.520 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # kill -0 71094 00:07:05.520 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@957 -- # uname 00:07:05.779 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:05.779 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71094 00:07:05.779 killing process with pid 71094 00:07:05.780 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:05.780 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:05.780 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71094' 00:07:05.780 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@971 -- # kill 71094 00:07:05.780 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@976 -- # wait 71094 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 71110 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@952 -- # '[' -z 71110 ']' 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # kill -0 71110 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@957 -- # uname 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71110 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:06.718 killing process with pid 71110 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71110' 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@971 -- # kill 71110 00:07:06.718 10:56:11 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@976 -- # wait 71110 00:07:06.977 ************************************ 00:07:06.977 END TEST locking_app_on_unlocked_coremask 00:07:06.978 ************************************ 00:07:06.978 00:07:06.978 real 0m3.537s 00:07:06.978 user 0m3.376s 00:07:06.978 sys 0m1.301s 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:06.978 10:56:12 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:07:06.978 10:56:12 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:06.978 10:56:12 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:06.978 10:56:12 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:06.978 ************************************ 00:07:06.978 START TEST locking_app_on_locked_coremask 00:07:06.978 ************************************ 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1127 -- # locking_app_on_locked_coremask 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=71172 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 71172 /var/tmp/spdk.sock 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71172 ']' 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:06.978 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:06.978 10:56:12 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:06.978 [2024-10-29 10:56:12.432720] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:06.978 [2024-10-29 10:56:12.432964] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71172 ] 00:07:07.237 [2024-10-29 10:56:12.604212] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.237 [2024-10-29 10:56:12.631179] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@866 -- # return 0 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=71184 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 71184 /var/tmp/spdk2.sock 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 71184 /var/tmp/spdk2.sock 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 71184 /var/tmp/spdk2.sock 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # '[' -z 71184 ']' 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:07.807 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:07.807 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:08.067 [2024-10-29 10:56:13.340496] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:08.067 [2024-10-29 10:56:13.340704] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71184 ] 00:07:08.067 [2024-10-29 10:56:13.503850] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 71172 has claimed it. 00:07:08.067 [2024-10-29 10:56:13.503936] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:07:08.636 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 848: kill: (71184) - No such process 00:07:08.636 ERROR: process (pid: 71184) is no longer running 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@866 -- # return 1 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 71172 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:07:08.636 10:56:13 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 71172 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 71172 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # '[' -z 71172 ']' 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # kill -0 71172 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # uname 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71172 00:07:08.895 killing process with pid 71172 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71172' 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@971 -- # kill 71172 00:07:08.895 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@976 -- # wait 71172 00:07:09.463 00:07:09.463 real 0m2.334s 00:07:09.463 user 0m2.516s 00:07:09.463 sys 0m0.687s 00:07:09.463 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:09.463 10:56:14 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:09.463 ************************************ 00:07:09.463 END TEST locking_app_on_locked_coremask 00:07:09.463 ************************************ 00:07:09.463 10:56:14 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:07:09.463 10:56:14 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:09.463 10:56:14 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:09.463 10:56:14 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:09.463 ************************************ 00:07:09.463 START TEST locking_overlapped_coremask 00:07:09.463 ************************************ 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1127 -- # locking_overlapped_coremask 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=71237 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 71237 /var/tmp/spdk.sock 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@833 -- # '[' -z 71237 ']' 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:09.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:09.463 10:56:14 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:09.463 [2024-10-29 10:56:14.832978] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:09.463 [2024-10-29 10:56:14.833238] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71237 ] 00:07:09.721 [2024-10-29 10:56:15.008190] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:09.721 [2024-10-29 10:56:15.037682] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:09.721 [2024-10-29 10:56:15.037791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.721 [2024-10-29 10:56:15.037880] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@866 -- # return 0 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=71255 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 71255 /var/tmp/spdk2.sock 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 71255 /var/tmp/spdk2.sock 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 71255 /var/tmp/spdk2.sock 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@833 -- # '[' -z 71255 ']' 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:10.288 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:10.288 10:56:15 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:10.288 [2024-10-29 10:56:15.721382] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:10.288 [2024-10-29 10:56:15.721951] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71255 ] 00:07:10.546 [2024-10-29 10:56:15.889102] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71237 has claimed it. 00:07:10.546 [2024-10-29 10:56:15.889183] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:07:11.114 ERROR: process (pid: 71255) is no longer running 00:07:11.114 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 848: kill: (71255) - No such process 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@866 -- # return 1 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 71237 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@952 -- # '[' -z 71237 ']' 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # kill -0 71237 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@957 -- # uname 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71237 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71237' 00:07:11.114 killing process with pid 71237 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@971 -- # kill 71237 00:07:11.114 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@976 -- # wait 71237 00:07:11.374 00:07:11.374 real 0m2.057s 00:07:11.374 user 0m5.467s 00:07:11.374 sys 0m0.530s 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:11.374 ************************************ 00:07:11.374 END TEST locking_overlapped_coremask 00:07:11.374 ************************************ 00:07:11.374 10:56:16 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:07:11.374 10:56:16 event.cpu_locks -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:11.374 10:56:16 event.cpu_locks -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:11.374 10:56:16 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:11.374 ************************************ 00:07:11.374 START TEST locking_overlapped_coremask_via_rpc 00:07:11.374 ************************************ 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1127 -- # locking_overlapped_coremask_via_rpc 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=71297 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 71297 /var/tmp/spdk.sock 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # '[' -z 71297 ']' 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:11.374 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:11.374 10:56:16 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:11.633 [2024-10-29 10:56:16.952512] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:11.633 [2024-10-29 10:56:16.952679] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71297 ] 00:07:11.633 [2024-10-29 10:56:17.103551] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:11.633 [2024-10-29 10:56:17.103609] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:11.633 [2024-10-29 10:56:17.131534] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:11.633 [2024-10-29 10:56:17.131654] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.633 [2024-10-29 10:56:17.131796] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@866 -- # return 0 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=71315 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 71315 /var/tmp/spdk2.sock 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # '[' -z 71315 ']' 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:12.584 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:12.584 10:56:17 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:12.584 [2024-10-29 10:56:17.847108] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:12.584 [2024-10-29 10:56:17.847330] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71315 ] 00:07:12.584 [2024-10-29 10:56:18.015788] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:12.584 [2024-10-29 10:56:18.015848] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:12.584 [2024-10-29 10:56:18.072950] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:07:12.584 [2024-10-29 10:56:18.072978] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:07:12.584 [2024-10-29 10:56:18.073073] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@866 -- # return 0 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.541 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:13.541 [2024-10-29 10:56:18.717550] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 71297 has claimed it. 00:07:13.541 request: 00:07:13.541 { 00:07:13.541 "method": "framework_enable_cpumask_locks", 00:07:13.541 "req_id": 1 00:07:13.541 } 00:07:13.541 Got JSON-RPC error response 00:07:13.541 response: 00:07:13.541 { 00:07:13.541 "code": -32603, 00:07:13.542 "message": "Failed to claim CPU core: 2" 00:07:13.542 } 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 71297 /var/tmp/spdk.sock 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # '[' -z 71297 ']' 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:13.542 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@866 -- # return 0 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 71315 /var/tmp/spdk2.sock 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # '[' -z 71315 ']' 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:13.542 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:13.542 10:56:18 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@866 -- # return 0 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:13.802 00:07:13.802 real 0m2.276s 00:07:13.802 user 0m1.039s 00:07:13.802 sys 0m0.170s 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:13.802 10:56:19 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:13.802 ************************************ 00:07:13.802 END TEST locking_overlapped_coremask_via_rpc 00:07:13.802 ************************************ 00:07:13.802 10:56:19 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:07:13.802 10:56:19 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71297 ]] 00:07:13.802 10:56:19 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71297 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@952 -- # '[' -z 71297 ']' 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@956 -- # kill -0 71297 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@957 -- # uname 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71297 00:07:13.802 killing process with pid 71297 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71297' 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@971 -- # kill 71297 00:07:13.802 10:56:19 event.cpu_locks -- common/autotest_common.sh@976 -- # wait 71297 00:07:14.372 10:56:19 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71315 ]] 00:07:14.372 10:56:19 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71315 00:07:14.372 10:56:19 event.cpu_locks -- common/autotest_common.sh@952 -- # '[' -z 71315 ']' 00:07:14.372 10:56:19 event.cpu_locks -- common/autotest_common.sh@956 -- # kill -0 71315 00:07:14.372 10:56:19 event.cpu_locks -- common/autotest_common.sh@957 -- # uname 00:07:14.372 10:56:19 event.cpu_locks -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:14.372 10:56:19 event.cpu_locks -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71315 00:07:14.630 killing process with pid 71315 00:07:14.630 10:56:19 event.cpu_locks -- common/autotest_common.sh@958 -- # process_name=reactor_2 00:07:14.630 10:56:19 event.cpu_locks -- common/autotest_common.sh@962 -- # '[' reactor_2 = sudo ']' 00:07:14.631 10:56:19 event.cpu_locks -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71315' 00:07:14.631 10:56:19 event.cpu_locks -- common/autotest_common.sh@971 -- # kill 71315 00:07:14.631 10:56:19 event.cpu_locks -- common/autotest_common.sh@976 -- # wait 71315 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 71297 ]] 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 71297 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@952 -- # '[' -z 71297 ']' 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@956 -- # kill -0 71297 00:07:14.888 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 956: kill: (71297) - No such process 00:07:14.888 Process with pid 71297 is not found 00:07:14.888 Process with pid 71315 is not found 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@979 -- # echo 'Process with pid 71297 is not found' 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 71315 ]] 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 71315 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@952 -- # '[' -z 71315 ']' 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@956 -- # kill -0 71315 00:07:14.888 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 956: kill: (71315) - No such process 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@979 -- # echo 'Process with pid 71315 is not found' 00:07:14.888 10:56:20 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:14.888 00:07:14.888 real 0m20.172s 00:07:14.888 user 0m32.515s 00:07:14.888 sys 0m6.657s 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:14.888 10:56:20 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:14.888 ************************************ 00:07:14.888 END TEST cpu_locks 00:07:14.888 ************************************ 00:07:14.888 00:07:14.888 real 0m46.644s 00:07:14.888 user 1m25.259s 00:07:14.888 sys 0m10.445s 00:07:14.888 10:56:20 event -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:14.888 10:56:20 event -- common/autotest_common.sh@10 -- # set +x 00:07:14.888 ************************************ 00:07:14.888 END TEST event 00:07:14.888 ************************************ 00:07:14.888 10:56:20 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:14.888 10:56:20 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:14.888 10:56:20 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:14.888 10:56:20 -- common/autotest_common.sh@10 -- # set +x 00:07:14.888 ************************************ 00:07:14.888 START TEST thread 00:07:14.888 ************************************ 00:07:14.888 10:56:20 thread -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:15.147 * Looking for test storage... 00:07:15.147 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1691 -- # lcov --version 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:15.147 10:56:20 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:15.147 10:56:20 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:15.147 10:56:20 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:15.147 10:56:20 thread -- scripts/common.sh@336 -- # IFS=.-: 00:07:15.147 10:56:20 thread -- scripts/common.sh@336 -- # read -ra ver1 00:07:15.147 10:56:20 thread -- scripts/common.sh@337 -- # IFS=.-: 00:07:15.147 10:56:20 thread -- scripts/common.sh@337 -- # read -ra ver2 00:07:15.147 10:56:20 thread -- scripts/common.sh@338 -- # local 'op=<' 00:07:15.147 10:56:20 thread -- scripts/common.sh@340 -- # ver1_l=2 00:07:15.147 10:56:20 thread -- scripts/common.sh@341 -- # ver2_l=1 00:07:15.147 10:56:20 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:15.147 10:56:20 thread -- scripts/common.sh@344 -- # case "$op" in 00:07:15.147 10:56:20 thread -- scripts/common.sh@345 -- # : 1 00:07:15.147 10:56:20 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:15.147 10:56:20 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:15.147 10:56:20 thread -- scripts/common.sh@365 -- # decimal 1 00:07:15.147 10:56:20 thread -- scripts/common.sh@353 -- # local d=1 00:07:15.147 10:56:20 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:15.147 10:56:20 thread -- scripts/common.sh@355 -- # echo 1 00:07:15.147 10:56:20 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:07:15.147 10:56:20 thread -- scripts/common.sh@366 -- # decimal 2 00:07:15.147 10:56:20 thread -- scripts/common.sh@353 -- # local d=2 00:07:15.147 10:56:20 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:15.147 10:56:20 thread -- scripts/common.sh@355 -- # echo 2 00:07:15.147 10:56:20 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:07:15.147 10:56:20 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:15.147 10:56:20 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:15.147 10:56:20 thread -- scripts/common.sh@368 -- # return 0 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:15.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.147 --rc genhtml_branch_coverage=1 00:07:15.147 --rc genhtml_function_coverage=1 00:07:15.147 --rc genhtml_legend=1 00:07:15.147 --rc geninfo_all_blocks=1 00:07:15.147 --rc geninfo_unexecuted_blocks=1 00:07:15.147 00:07:15.147 ' 00:07:15.147 10:56:20 thread -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:15.147 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.147 --rc genhtml_branch_coverage=1 00:07:15.147 --rc genhtml_function_coverage=1 00:07:15.147 --rc genhtml_legend=1 00:07:15.147 --rc geninfo_all_blocks=1 00:07:15.148 --rc geninfo_unexecuted_blocks=1 00:07:15.148 00:07:15.148 ' 00:07:15.148 10:56:20 thread -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:15.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.148 --rc genhtml_branch_coverage=1 00:07:15.148 --rc genhtml_function_coverage=1 00:07:15.148 --rc genhtml_legend=1 00:07:15.148 --rc geninfo_all_blocks=1 00:07:15.148 --rc geninfo_unexecuted_blocks=1 00:07:15.148 00:07:15.148 ' 00:07:15.148 10:56:20 thread -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:15.148 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:15.148 --rc genhtml_branch_coverage=1 00:07:15.148 --rc genhtml_function_coverage=1 00:07:15.148 --rc genhtml_legend=1 00:07:15.148 --rc geninfo_all_blocks=1 00:07:15.148 --rc geninfo_unexecuted_blocks=1 00:07:15.148 00:07:15.148 ' 00:07:15.148 10:56:20 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:15.148 10:56:20 thread -- common/autotest_common.sh@1103 -- # '[' 8 -le 1 ']' 00:07:15.148 10:56:20 thread -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:15.148 10:56:20 thread -- common/autotest_common.sh@10 -- # set +x 00:07:15.148 ************************************ 00:07:15.148 START TEST thread_poller_perf 00:07:15.148 ************************************ 00:07:15.148 10:56:20 thread.thread_poller_perf -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:15.408 [2024-10-29 10:56:20.651462] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:15.408 [2024-10-29 10:56:20.651597] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71444 ] 00:07:15.408 [2024-10-29 10:56:20.821292] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:15.408 [2024-10-29 10:56:20.863844] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.408 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:07:16.785 [2024-10-29T10:56:22.283Z] ====================================== 00:07:16.785 [2024-10-29T10:56:22.283Z] busy:2299313990 (cyc) 00:07:16.785 [2024-10-29T10:56:22.283Z] total_run_count: 385000 00:07:16.785 [2024-10-29T10:56:22.283Z] tsc_hz: 2290000000 (cyc) 00:07:16.785 [2024-10-29T10:56:22.283Z] ====================================== 00:07:16.785 [2024-10-29T10:56:22.283Z] poller_cost: 5972 (cyc), 2607 (nsec) 00:07:16.785 00:07:16.785 real 0m1.354s 00:07:16.785 user 0m1.143s 00:07:16.785 sys 0m0.103s 00:07:16.785 10:56:21 thread.thread_poller_perf -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:16.785 10:56:21 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:16.785 ************************************ 00:07:16.785 END TEST thread_poller_perf 00:07:16.785 ************************************ 00:07:16.785 10:56:22 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:16.785 10:56:22 thread -- common/autotest_common.sh@1103 -- # '[' 8 -le 1 ']' 00:07:16.785 10:56:22 thread -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:16.785 10:56:22 thread -- common/autotest_common.sh@10 -- # set +x 00:07:16.785 ************************************ 00:07:16.785 START TEST thread_poller_perf 00:07:16.785 ************************************ 00:07:16.785 10:56:22 thread.thread_poller_perf -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:16.785 [2024-10-29 10:56:22.080661] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:16.785 [2024-10-29 10:56:22.080803] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71486 ] 00:07:16.785 [2024-10-29 10:56:22.259076] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:17.045 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:07:17.045 [2024-10-29 10:56:22.304231] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.984 [2024-10-29T10:56:23.482Z] ====================================== 00:07:17.984 [2024-10-29T10:56:23.482Z] busy:2293780838 (cyc) 00:07:17.984 [2024-10-29T10:56:23.482Z] total_run_count: 4998000 00:07:17.984 [2024-10-29T10:56:23.482Z] tsc_hz: 2290000000 (cyc) 00:07:17.984 [2024-10-29T10:56:23.482Z] ====================================== 00:07:17.984 [2024-10-29T10:56:23.482Z] poller_cost: 458 (cyc), 200 (nsec) 00:07:17.984 00:07:17.984 real 0m1.361s 00:07:17.984 user 0m1.144s 00:07:17.984 sys 0m0.108s 00:07:17.984 ************************************ 00:07:17.984 END TEST thread_poller_perf 00:07:17.984 ************************************ 00:07:17.984 10:56:23 thread.thread_poller_perf -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:17.984 10:56:23 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:17.984 10:56:23 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:07:17.984 ************************************ 00:07:17.984 END TEST thread 00:07:17.984 ************************************ 00:07:17.984 00:07:17.984 real 0m3.061s 00:07:17.984 user 0m2.446s 00:07:17.984 sys 0m0.417s 00:07:17.984 10:56:23 thread -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:17.984 10:56:23 thread -- common/autotest_common.sh@10 -- # set +x 00:07:18.244 10:56:23 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:07:18.244 10:56:23 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:18.244 10:56:23 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:18.244 10:56:23 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:18.244 10:56:23 -- common/autotest_common.sh@10 -- # set +x 00:07:18.244 ************************************ 00:07:18.244 START TEST app_cmdline 00:07:18.244 ************************************ 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:18.244 * Looking for test storage... 00:07:18.244 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1691 -- # lcov --version 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@345 -- # : 1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:18.244 10:56:23 app_cmdline -- scripts/common.sh@368 -- # return 0 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:18.244 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:18.244 --rc genhtml_branch_coverage=1 00:07:18.244 --rc genhtml_function_coverage=1 00:07:18.244 --rc genhtml_legend=1 00:07:18.244 --rc geninfo_all_blocks=1 00:07:18.244 --rc geninfo_unexecuted_blocks=1 00:07:18.244 00:07:18.244 ' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:18.244 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:18.244 --rc genhtml_branch_coverage=1 00:07:18.244 --rc genhtml_function_coverage=1 00:07:18.244 --rc genhtml_legend=1 00:07:18.244 --rc geninfo_all_blocks=1 00:07:18.244 --rc geninfo_unexecuted_blocks=1 00:07:18.244 00:07:18.244 ' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:18.244 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:18.244 --rc genhtml_branch_coverage=1 00:07:18.244 --rc genhtml_function_coverage=1 00:07:18.244 --rc genhtml_legend=1 00:07:18.244 --rc geninfo_all_blocks=1 00:07:18.244 --rc geninfo_unexecuted_blocks=1 00:07:18.244 00:07:18.244 ' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:18.244 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:18.244 --rc genhtml_branch_coverage=1 00:07:18.244 --rc genhtml_function_coverage=1 00:07:18.244 --rc genhtml_legend=1 00:07:18.244 --rc geninfo_all_blocks=1 00:07:18.244 --rc geninfo_unexecuted_blocks=1 00:07:18.244 00:07:18.244 ' 00:07:18.244 10:56:23 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:07:18.244 10:56:23 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71564 00:07:18.244 10:56:23 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:07:18.244 10:56:23 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71564 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@833 -- # '[' -z 71564 ']' 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:18.244 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:18.244 10:56:23 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:18.504 [2024-10-29 10:56:23.822399] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:18.504 [2024-10-29 10:56:23.822604] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71564 ] 00:07:18.504 [2024-10-29 10:56:23.991770] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:18.763 [2024-10-29 10:56:24.032799] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:19.333 10:56:24 app_cmdline -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:19.333 10:56:24 app_cmdline -- common/autotest_common.sh@866 -- # return 0 00:07:19.333 10:56:24 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:07:19.333 { 00:07:19.333 "version": "SPDK v25.01-pre git sha1 12fc2abf1", 00:07:19.333 "fields": { 00:07:19.333 "major": 25, 00:07:19.333 "minor": 1, 00:07:19.333 "patch": 0, 00:07:19.333 "suffix": "-pre", 00:07:19.333 "commit": "12fc2abf1" 00:07:19.333 } 00:07:19.333 } 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@26 -- # sort 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:07:19.593 10:56:24 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:07:19.593 10:56:24 app_cmdline -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:19.593 request: 00:07:19.593 { 00:07:19.593 "method": "env_dpdk_get_mem_stats", 00:07:19.593 "req_id": 1 00:07:19.593 } 00:07:19.593 Got JSON-RPC error response 00:07:19.593 response: 00:07:19.593 { 00:07:19.593 "code": -32601, 00:07:19.593 "message": "Method not found" 00:07:19.593 } 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:19.853 10:56:25 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71564 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@952 -- # '[' -z 71564 ']' 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@956 -- # kill -0 71564 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@957 -- # uname 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71564 00:07:19.853 killing process with pid 71564 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71564' 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@971 -- # kill 71564 00:07:19.853 10:56:25 app_cmdline -- common/autotest_common.sh@976 -- # wait 71564 00:07:20.113 ************************************ 00:07:20.113 END TEST app_cmdline 00:07:20.113 ************************************ 00:07:20.113 00:07:20.113 real 0m1.996s 00:07:20.113 user 0m2.200s 00:07:20.113 sys 0m0.579s 00:07:20.113 10:56:25 app_cmdline -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:20.113 10:56:25 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:20.113 10:56:25 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:20.113 10:56:25 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:20.113 10:56:25 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:20.113 10:56:25 -- common/autotest_common.sh@10 -- # set +x 00:07:20.113 ************************************ 00:07:20.113 START TEST version 00:07:20.113 ************************************ 00:07:20.113 10:56:25 version -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:20.373 * Looking for test storage... 00:07:20.373 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:20.373 10:56:25 version -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:20.373 10:56:25 version -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:20.373 10:56:25 version -- common/autotest_common.sh@1691 -- # lcov --version 00:07:20.373 10:56:25 version -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:20.373 10:56:25 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:20.373 10:56:25 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:20.373 10:56:25 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:20.373 10:56:25 version -- scripts/common.sh@336 -- # IFS=.-: 00:07:20.373 10:56:25 version -- scripts/common.sh@336 -- # read -ra ver1 00:07:20.373 10:56:25 version -- scripts/common.sh@337 -- # IFS=.-: 00:07:20.374 10:56:25 version -- scripts/common.sh@337 -- # read -ra ver2 00:07:20.374 10:56:25 version -- scripts/common.sh@338 -- # local 'op=<' 00:07:20.374 10:56:25 version -- scripts/common.sh@340 -- # ver1_l=2 00:07:20.374 10:56:25 version -- scripts/common.sh@341 -- # ver2_l=1 00:07:20.374 10:56:25 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:20.374 10:56:25 version -- scripts/common.sh@344 -- # case "$op" in 00:07:20.374 10:56:25 version -- scripts/common.sh@345 -- # : 1 00:07:20.374 10:56:25 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:20.374 10:56:25 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:20.374 10:56:25 version -- scripts/common.sh@365 -- # decimal 1 00:07:20.374 10:56:25 version -- scripts/common.sh@353 -- # local d=1 00:07:20.374 10:56:25 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:20.374 10:56:25 version -- scripts/common.sh@355 -- # echo 1 00:07:20.374 10:56:25 version -- scripts/common.sh@365 -- # ver1[v]=1 00:07:20.374 10:56:25 version -- scripts/common.sh@366 -- # decimal 2 00:07:20.374 10:56:25 version -- scripts/common.sh@353 -- # local d=2 00:07:20.374 10:56:25 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:20.374 10:56:25 version -- scripts/common.sh@355 -- # echo 2 00:07:20.374 10:56:25 version -- scripts/common.sh@366 -- # ver2[v]=2 00:07:20.374 10:56:25 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:20.374 10:56:25 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:20.374 10:56:25 version -- scripts/common.sh@368 -- # return 0 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:20.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.374 --rc genhtml_branch_coverage=1 00:07:20.374 --rc genhtml_function_coverage=1 00:07:20.374 --rc genhtml_legend=1 00:07:20.374 --rc geninfo_all_blocks=1 00:07:20.374 --rc geninfo_unexecuted_blocks=1 00:07:20.374 00:07:20.374 ' 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:20.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.374 --rc genhtml_branch_coverage=1 00:07:20.374 --rc genhtml_function_coverage=1 00:07:20.374 --rc genhtml_legend=1 00:07:20.374 --rc geninfo_all_blocks=1 00:07:20.374 --rc geninfo_unexecuted_blocks=1 00:07:20.374 00:07:20.374 ' 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:20.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.374 --rc genhtml_branch_coverage=1 00:07:20.374 --rc genhtml_function_coverage=1 00:07:20.374 --rc genhtml_legend=1 00:07:20.374 --rc geninfo_all_blocks=1 00:07:20.374 --rc geninfo_unexecuted_blocks=1 00:07:20.374 00:07:20.374 ' 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:20.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.374 --rc genhtml_branch_coverage=1 00:07:20.374 --rc genhtml_function_coverage=1 00:07:20.374 --rc genhtml_legend=1 00:07:20.374 --rc geninfo_all_blocks=1 00:07:20.374 --rc geninfo_unexecuted_blocks=1 00:07:20.374 00:07:20.374 ' 00:07:20.374 10:56:25 version -- app/version.sh@17 -- # get_header_version major 00:07:20.374 10:56:25 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # cut -f2 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # tr -d '"' 00:07:20.374 10:56:25 version -- app/version.sh@17 -- # major=25 00:07:20.374 10:56:25 version -- app/version.sh@18 -- # get_header_version minor 00:07:20.374 10:56:25 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # cut -f2 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # tr -d '"' 00:07:20.374 10:56:25 version -- app/version.sh@18 -- # minor=1 00:07:20.374 10:56:25 version -- app/version.sh@19 -- # get_header_version patch 00:07:20.374 10:56:25 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # cut -f2 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # tr -d '"' 00:07:20.374 10:56:25 version -- app/version.sh@19 -- # patch=0 00:07:20.374 10:56:25 version -- app/version.sh@20 -- # get_header_version suffix 00:07:20.374 10:56:25 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # cut -f2 00:07:20.374 10:56:25 version -- app/version.sh@14 -- # tr -d '"' 00:07:20.374 10:56:25 version -- app/version.sh@20 -- # suffix=-pre 00:07:20.374 10:56:25 version -- app/version.sh@22 -- # version=25.1 00:07:20.374 10:56:25 version -- app/version.sh@25 -- # (( patch != 0 )) 00:07:20.374 10:56:25 version -- app/version.sh@28 -- # version=25.1rc0 00:07:20.374 10:56:25 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:07:20.374 10:56:25 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:07:20.374 10:56:25 version -- app/version.sh@30 -- # py_version=25.1rc0 00:07:20.374 10:56:25 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:07:20.374 ************************************ 00:07:20.374 END TEST version 00:07:20.374 ************************************ 00:07:20.374 00:07:20.374 real 0m0.301s 00:07:20.374 user 0m0.184s 00:07:20.374 sys 0m0.174s 00:07:20.374 10:56:25 version -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:20.374 10:56:25 version -- common/autotest_common.sh@10 -- # set +x 00:07:20.634 10:56:25 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:07:20.634 10:56:25 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:07:20.634 10:56:25 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:20.634 10:56:25 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:20.634 10:56:25 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:20.634 10:56:25 -- common/autotest_common.sh@10 -- # set +x 00:07:20.634 ************************************ 00:07:20.634 START TEST bdev_raid 00:07:20.634 ************************************ 00:07:20.634 10:56:25 bdev_raid -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:20.634 * Looking for test storage... 00:07:20.634 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:07:20.634 10:56:26 bdev_raid -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:07:20.634 10:56:26 bdev_raid -- common/autotest_common.sh@1691 -- # lcov --version 00:07:20.634 10:56:26 bdev_raid -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:07:20.634 10:56:26 bdev_raid -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@345 -- # : 1 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:20.634 10:56:26 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:20.893 10:56:26 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:07:20.893 10:56:26 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:20.894 10:56:26 bdev_raid -- scripts/common.sh@368 -- # return 0 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:07:20.894 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.894 --rc genhtml_branch_coverage=1 00:07:20.894 --rc genhtml_function_coverage=1 00:07:20.894 --rc genhtml_legend=1 00:07:20.894 --rc geninfo_all_blocks=1 00:07:20.894 --rc geninfo_unexecuted_blocks=1 00:07:20.894 00:07:20.894 ' 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:07:20.894 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.894 --rc genhtml_branch_coverage=1 00:07:20.894 --rc genhtml_function_coverage=1 00:07:20.894 --rc genhtml_legend=1 00:07:20.894 --rc geninfo_all_blocks=1 00:07:20.894 --rc geninfo_unexecuted_blocks=1 00:07:20.894 00:07:20.894 ' 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:07:20.894 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.894 --rc genhtml_branch_coverage=1 00:07:20.894 --rc genhtml_function_coverage=1 00:07:20.894 --rc genhtml_legend=1 00:07:20.894 --rc geninfo_all_blocks=1 00:07:20.894 --rc geninfo_unexecuted_blocks=1 00:07:20.894 00:07:20.894 ' 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:07:20.894 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:20.894 --rc genhtml_branch_coverage=1 00:07:20.894 --rc genhtml_function_coverage=1 00:07:20.894 --rc genhtml_legend=1 00:07:20.894 --rc geninfo_all_blocks=1 00:07:20.894 --rc geninfo_unexecuted_blocks=1 00:07:20.894 00:07:20.894 ' 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:20.894 10:56:26 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:07:20.894 10:56:26 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:20.894 10:56:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:20.894 ************************************ 00:07:20.894 START TEST raid1_resize_data_offset_test 00:07:20.894 ************************************ 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1127 -- # raid_resize_data_offset_test 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71735 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71735' 00:07:20.894 Process raid pid: 71735 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71735 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@833 -- # '[' -z 71735 ']' 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:20.894 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:20.894 10:56:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.894 [2024-10-29 10:56:26.252053] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:20.894 [2024-10-29 10:56:26.252608] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:21.154 [2024-10-29 10:56:26.403450] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:21.154 [2024-10-29 10:56:26.428525] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:21.154 [2024-10-29 10:56:26.471051] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:21.154 [2024-10-29 10:56:26.471185] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@866 -- # return 0 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 malloc0 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 malloc1 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 null0 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 [2024-10-29 10:56:27.150240] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:07:21.724 [2024-10-29 10:56:27.152150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:21.724 [2024-10-29 10:56:27.152195] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:07:21.724 [2024-10-29 10:56:27.152339] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:21.724 [2024-10-29 10:56:27.152350] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:07:21.724 [2024-10-29 10:56:27.152706] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:21.724 [2024-10-29 10:56:27.152880] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:21.724 [2024-10-29 10:56:27.152895] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:21.724 [2024-10-29 10:56:27.153048] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.724 [2024-10-29 10:56:27.210117] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.724 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.984 malloc2 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.984 [2024-10-29 10:56:27.334667] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:21.984 [2024-10-29 10:56:27.339976] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.984 [2024-10-29 10:56:27.341785] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71735 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@952 -- # '[' -z 71735 ']' 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # kill -0 71735 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@957 -- # uname 00:07:21.984 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71735 00:07:21.985 killing process with pid 71735 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71735' 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@971 -- # kill 71735 00:07:21.985 [2024-10-29 10:56:27.428765] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:21.985 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@976 -- # wait 71735 00:07:21.985 [2024-10-29 10:56:27.429764] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:07:21.985 [2024-10-29 10:56:27.429893] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.985 [2024-10-29 10:56:27.429913] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:07:21.985 [2024-10-29 10:56:27.435743] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:21.985 [2024-10-29 10:56:27.436022] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:21.985 [2024-10-29 10:56:27.436037] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:22.244 [2024-10-29 10:56:27.645798] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:22.504 10:56:27 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:07:22.504 00:07:22.504 real 0m1.684s 00:07:22.504 user 0m1.692s 00:07:22.504 sys 0m0.429s 00:07:22.504 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:22.504 10:56:27 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.504 ************************************ 00:07:22.504 END TEST raid1_resize_data_offset_test 00:07:22.504 ************************************ 00:07:22.504 10:56:27 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:07:22.504 10:56:27 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:22.504 10:56:27 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:22.504 10:56:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:22.504 ************************************ 00:07:22.504 START TEST raid0_resize_superblock_test 00:07:22.504 ************************************ 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1127 -- # raid_resize_superblock_test 0 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71786 00:07:22.504 Process raid pid: 71786 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71786' 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71786 00:07:22.504 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 71786 ']' 00:07:22.505 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:22.505 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:22.505 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:22.505 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:22.505 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:22.505 10:56:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:22.764 [2024-10-29 10:56:28.009161] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:22.764 [2024-10-29 10:56:28.009383] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:22.764 [2024-10-29 10:56:28.161462] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:22.764 [2024-10-29 10:56:28.188905] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:22.764 [2024-10-29 10:56:28.231710] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:22.764 [2024-10-29 10:56:28.231841] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 malloc0 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 [2024-10-29 10:56:28.972377] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:23.785 [2024-10-29 10:56:28.972446] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:23.785 [2024-10-29 10:56:28.972479] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:23.785 [2024-10-29 10:56:28.972490] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:23.785 [2024-10-29 10:56:28.974540] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:23.785 [2024-10-29 10:56:28.974580] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:23.785 pt0 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 f2c6d266-c3de-4be7-b45f-460b1ef7a011 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 fd653ae7-ec65-47e2-a08d-d9cc18b175e3 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 3309352b-ead9-4046-a6db-9011463eb76f 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 [2024-10-29 10:56:29.108132] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev fd653ae7-ec65-47e2-a08d-d9cc18b175e3 is claimed 00:07:23.785 [2024-10-29 10:56:29.108309] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3309352b-ead9-4046-a6db-9011463eb76f is claimed 00:07:23.785 [2024-10-29 10:56:29.108475] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:23.785 [2024-10-29 10:56:29.108492] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:07:23.785 [2024-10-29 10:56:29.108732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:23.785 [2024-10-29 10:56:29.108877] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:23.785 [2024-10-29 10:56:29.108892] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:23.785 [2024-10-29 10:56:29.109011] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.785 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.785 [2024-10-29 10:56:29.220121] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.786 [2024-10-29 10:56:29.248003] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:23.786 [2024-10-29 10:56:29.248033] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'fd653ae7-ec65-47e2-a08d-d9cc18b175e3' was resized: old size 131072, new size 204800 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.786 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.786 [2024-10-29 10:56:29.259921] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:23.786 [2024-10-29 10:56:29.259943] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '3309352b-ead9-4046-a6db-9011463eb76f' was resized: old size 131072, new size 204800 00:07:23.786 [2024-10-29 10:56:29.259971] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 [2024-10-29 10:56:29.371826] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 [2024-10-29 10:56:29.415556] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:24.046 [2024-10-29 10:56:29.415629] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:24.046 [2024-10-29 10:56:29.415640] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:24.046 [2024-10-29 10:56:29.415654] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:24.046 [2024-10-29 10:56:29.415741] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:24.046 [2024-10-29 10:56:29.415780] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:24.046 [2024-10-29 10:56:29.415792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 [2024-10-29 10:56:29.427517] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:24.046 [2024-10-29 10:56:29.427569] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.046 [2024-10-29 10:56:29.427589] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:07:24.046 [2024-10-29 10:56:29.427599] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.046 [2024-10-29 10:56:29.429675] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.046 [2024-10-29 10:56:29.429721] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:24.046 [2024-10-29 10:56:29.431091] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev fd653ae7-ec65-47e2-a08d-d9cc18b175e3 00:07:24.046 [2024-10-29 10:56:29.431187] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev fd653ae7-ec65-47e2-a08d-d9cc18b175e3 is claimed 00:07:24.046 [2024-10-29 10:56:29.431280] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 3309352b-ead9-4046-a6db-9011463eb76f 00:07:24.046 [2024-10-29 10:56:29.431301] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 3309352b-ead9-4046-a6db-9011463eb76f is claimed 00:07:24.046 [2024-10-29 10:56:29.431413] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 3309352b-ead9-4046-a6db-9011463eb76f (2) smaller than existing raid bdev Raid (3) 00:07:24.046 [2024-10-29 10:56:29.431435] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev fd653ae7-ec65-47e2-a08d-d9cc18b175e3: File exists 00:07:24.046 [2024-10-29 10:56:29.431471] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:07:24.046 [2024-10-29 10:56:29.431480] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:07:24.046 [2024-10-29 10:56:29.431695] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:24.046 [2024-10-29 10:56:29.431808] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:07:24.046 [2024-10-29 10:56:29.431815] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006600 00:07:24.046 [2024-10-29 10:56:29.431953] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:24.046 pt0 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.046 [2024-10-29 10:56:29.456111] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:07:24.046 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71786 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 71786 ']' 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # kill -0 71786 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@957 -- # uname 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71786 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:24.047 killing process with pid 71786 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71786' 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@971 -- # kill 71786 00:07:24.047 [2024-10-29 10:56:29.535044] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:24.047 [2024-10-29 10:56:29.535116] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:24.047 [2024-10-29 10:56:29.535156] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:24.047 [2024-10-29 10:56:29.535165] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Raid, state offline 00:07:24.047 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@976 -- # wait 71786 00:07:24.306 [2024-10-29 10:56:29.694753] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:24.566 10:56:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:24.566 00:07:24.566 real 0m1.978s 00:07:24.566 user 0m2.261s 00:07:24.566 sys 0m0.491s 00:07:24.566 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:24.566 10:56:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.566 ************************************ 00:07:24.566 END TEST raid0_resize_superblock_test 00:07:24.566 ************************************ 00:07:24.566 10:56:29 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:07:24.566 10:56:29 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:24.566 10:56:29 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:24.566 10:56:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:24.566 ************************************ 00:07:24.566 START TEST raid1_resize_superblock_test 00:07:24.566 ************************************ 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1127 -- # raid_resize_superblock_test 1 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71857 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71857' 00:07:24.566 Process raid pid: 71857 00:07:24.566 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71857 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 71857 ']' 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:24.566 10:56:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.825 [2024-10-29 10:56:30.076617] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:24.825 [2024-10-29 10:56:30.076866] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:24.825 [2024-10-29 10:56:30.238387] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:24.825 [2024-10-29 10:56:30.263194] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:24.825 [2024-10-29 10:56:30.305669] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:24.825 [2024-10-29 10:56:30.305784] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.765 10:56:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:25.765 10:56:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:07:25.765 10:56:30 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:25.765 10:56:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 malloc0 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 [2024-10-29 10:56:31.037513] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:25.765 [2024-10-29 10:56:31.037636] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.765 [2024-10-29 10:56:31.037679] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:25.765 [2024-10-29 10:56:31.037711] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.765 [2024-10-29 10:56:31.039873] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.765 [2024-10-29 10:56:31.039948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:25.765 pt0 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 ed556fda-e887-4450-b353-7ba19f12603c 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 f90b00ba-55c1-4a33-9d7a-2eeac30532d8 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 6ce17c12-1cbb-44ee-8811-8dbe006246a6 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 [2024-10-29 10:56:31.172839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev f90b00ba-55c1-4a33-9d7a-2eeac30532d8 is claimed 00:07:25.765 [2024-10-29 10:56:31.172952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 6ce17c12-1cbb-44ee-8811-8dbe006246a6 is claimed 00:07:25.765 [2024-10-29 10:56:31.173088] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:25.765 [2024-10-29 10:56:31.173101] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:07:25.765 [2024-10-29 10:56:31.173369] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:25.765 [2024-10-29 10:56:31.173553] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:25.765 [2024-10-29 10:56:31.173567] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:25.765 [2024-10-29 10:56:31.173714] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.765 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 [2024-10-29 10:56:31.269001] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 [2024-10-29 10:56:31.316747] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:26.026 [2024-10-29 10:56:31.316776] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'f90b00ba-55c1-4a33-9d7a-2eeac30532d8' was resized: old size 131072, new size 204800 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 [2024-10-29 10:56:31.328631] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:26.026 [2024-10-29 10:56:31.328710] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '6ce17c12-1cbb-44ee-8811-8dbe006246a6' was resized: old size 131072, new size 204800 00:07:26.026 [2024-10-29 10:56:31.328745] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 [2024-10-29 10:56:31.440518] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.026 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.026 [2024-10-29 10:56:31.472252] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:26.026 [2024-10-29 10:56:31.472368] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:26.026 [2024-10-29 10:56:31.472402] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:26.027 [2024-10-29 10:56:31.472580] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:26.027 [2024-10-29 10:56:31.472731] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.027 [2024-10-29 10:56:31.472788] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.027 [2024-10-29 10:56:31.472801] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.027 [2024-10-29 10:56:31.484208] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:26.027 [2024-10-29 10:56:31.484262] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:26.027 [2024-10-29 10:56:31.484298] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:07:26.027 [2024-10-29 10:56:31.484308] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:26.027 [2024-10-29 10:56:31.486429] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:26.027 [2024-10-29 10:56:31.486515] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:26.027 [2024-10-29 10:56:31.487930] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev f90b00ba-55c1-4a33-9d7a-2eeac30532d8 00:07:26.027 [2024-10-29 10:56:31.487988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev f90b00ba-55c1-4a33-9d7a-2eeac30532d8 is claimed 00:07:26.027 [2024-10-29 10:56:31.488063] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 6ce17c12-1cbb-44ee-8811-8dbe006246a6 00:07:26.027 [2024-10-29 10:56:31.488082] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 6ce17c12-1cbb-44ee-8811-8dbe006246a6 is claimed 00:07:26.027 [2024-10-29 10:56:31.488165] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 6ce17c12-1cbb-44ee-8811-8dbe006246a6 (2) smaller than existing raid bdev Raid (3) 00:07:26.027 [2024-10-29 10:56:31.488182] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev f90b00ba-55c1-4a33-9d7a-2eeac30532d8: File exists 00:07:26.027 [2024-10-29 10:56:31.488224] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:07:26.027 [2024-10-29 10:56:31.488232] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:26.027 [2024-10-29 10:56:31.488470] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:07:26.027 [2024-10-29 10:56:31.488610] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:07:26.027 [2024-10-29 10:56:31.488619] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006600 00:07:26.027 [2024-10-29 10:56:31.488759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:26.027 pt0 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:07:26.027 [2024-10-29 10:56:31.508443] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.027 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71857 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 71857 ']' 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # kill -0 71857 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@957 -- # uname 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71857 00:07:26.287 killing process with pid 71857 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71857' 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@971 -- # kill 71857 00:07:26.287 [2024-10-29 10:56:31.577410] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:26.287 [2024-10-29 10:56:31.577468] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.287 [2024-10-29 10:56:31.577512] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.287 [2024-10-29 10:56:31.577520] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Raid, state offline 00:07:26.287 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@976 -- # wait 71857 00:07:26.287 [2024-10-29 10:56:31.736908] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:26.547 ************************************ 00:07:26.547 END TEST raid1_resize_superblock_test 00:07:26.547 ************************************ 00:07:26.547 10:56:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:26.547 00:07:26.547 real 0m1.954s 00:07:26.547 user 0m2.270s 00:07:26.547 sys 0m0.442s 00:07:26.547 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:26.547 10:56:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:07:26.547 10:56:31 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:07:26.547 10:56:31 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:26.547 10:56:31 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:26.547 10:56:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:26.547 ************************************ 00:07:26.547 START TEST raid_function_test_raid0 00:07:26.547 ************************************ 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1127 -- # raid_function_test raid0 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71932 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71932' 00:07:26.547 Process raid pid: 71932 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71932 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@833 -- # '[' -z 71932 ']' 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:26.547 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:26.547 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:26.808 [2024-10-29 10:56:32.091963] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:26.808 [2024-10-29 10:56:32.092178] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:26.808 [2024-10-29 10:56:32.262218] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:26.808 [2024-10-29 10:56:32.287468] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:27.068 [2024-10-29 10:56:32.330718] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.068 [2024-10-29 10:56:32.330801] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@866 -- # return 0 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:27.638 Base_1 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:27.638 Base_2 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.638 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:27.638 [2024-10-29 10:56:32.942113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:27.639 [2024-10-29 10:56:32.944006] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:27.639 [2024-10-29 10:56:32.944109] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:27.639 [2024-10-29 10:56:32.944187] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:27.639 [2024-10-29 10:56:32.944519] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:27.639 [2024-10-29 10:56:32.944685] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:27.639 [2024-10-29 10:56:32.944726] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000006280 00:07:27.639 [2024-10-29 10:56:32.944916] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:27.639 10:56:32 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:27.898 [2024-10-29 10:56:33.169807] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:27.898 /dev/nbd0 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # local i 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # break 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:27.898 1+0 records in 00:07:27.898 1+0 records out 00:07:27.898 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000232262 s, 17.6 MB/s 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # size=4096 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # return 0 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:27.898 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:27.899 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:27.899 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:28.158 { 00:07:28.158 "nbd_device": "/dev/nbd0", 00:07:28.158 "bdev_name": "raid" 00:07:28.158 } 00:07:28.158 ]' 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:28.158 { 00:07:28.158 "nbd_device": "/dev/nbd0", 00:07:28.158 "bdev_name": "raid" 00:07:28.158 } 00:07:28.158 ]' 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:28.158 4096+0 records in 00:07:28.158 4096+0 records out 00:07:28.158 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0218187 s, 96.1 MB/s 00:07:28.158 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:28.419 4096+0 records in 00:07:28.419 4096+0 records out 00:07:28.419 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.220229 s, 9.5 MB/s 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:28.419 128+0 records in 00:07:28.419 128+0 records out 00:07:28.419 65536 bytes (66 kB, 64 KiB) copied, 0.00119201 s, 55.0 MB/s 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:28.419 2035+0 records in 00:07:28.419 2035+0 records out 00:07:28.419 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0154533 s, 67.4 MB/s 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:28.419 456+0 records in 00:07:28.419 456+0 records out 00:07:28.419 233472 bytes (233 kB, 228 KiB) copied, 0.00261256 s, 89.4 MB/s 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:28.419 10:56:33 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:28.679 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:28.679 [2024-10-29 10:56:34.063258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:28.679 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:28.679 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:28.679 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:28.680 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71932 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@952 -- # '[' -z 71932 ']' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # kill -0 71932 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@957 -- # uname 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 71932 00:07:28.940 killing process with pid 71932 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@970 -- # echo 'killing process with pid 71932' 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@971 -- # kill 71932 00:07:28.940 [2024-10-29 10:56:34.360880] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:28.940 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@976 -- # wait 71932 00:07:28.940 [2024-10-29 10:56:34.361050] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:28.940 [2024-10-29 10:56:34.361113] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:28.940 [2024-10-29 10:56:34.361126] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid, state offline 00:07:28.940 [2024-10-29 10:56:34.404786] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:29.512 ************************************ 00:07:29.512 END TEST raid_function_test_raid0 00:07:29.512 ************************************ 00:07:29.512 10:56:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:07:29.512 00:07:29.512 real 0m2.731s 00:07:29.512 user 0m3.295s 00:07:29.512 sys 0m0.882s 00:07:29.512 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:29.512 10:56:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:29.512 10:56:34 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:07:29.512 10:56:34 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:29.512 10:56:34 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:29.512 10:56:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:29.512 ************************************ 00:07:29.512 START TEST raid_function_test_concat 00:07:29.512 ************************************ 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1127 -- # raid_function_test concat 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=72046 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 72046' 00:07:29.512 Process raid pid: 72046 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 72046 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@833 -- # '[' -z 72046 ']' 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:29.512 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:29.512 10:56:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:29.512 [2024-10-29 10:56:34.889694] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:29.512 [2024-10-29 10:56:34.889898] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:29.772 [2024-10-29 10:56:35.039439] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:29.772 [2024-10-29 10:56:35.080786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:29.772 [2024-10-29 10:56:35.157169] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:29.772 [2024-10-29 10:56:35.157284] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@866 -- # return 0 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 Base_1 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 Base_2 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 [2024-10-29 10:56:35.763415] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:30.343 [2024-10-29 10:56:35.765660] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:30.343 [2024-10-29 10:56:35.765723] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:30.343 [2024-10-29 10:56:35.765744] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:30.343 [2024-10-29 10:56:35.766017] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:30.343 [2024-10-29 10:56:35.766170] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:30.343 [2024-10-29 10:56:35.766180] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000006280 00:07:30.343 [2024-10-29 10:56:35.766334] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:30.343 10:56:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:30.603 [2024-10-29 10:56:36.003087] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:30.603 /dev/nbd0 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # local i 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # break 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:30.603 1+0 records in 00:07:30.603 1+0 records out 00:07:30.603 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000347565 s, 11.8 MB/s 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # size=4096 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # return 0 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:30.603 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:30.863 { 00:07:30.863 "nbd_device": "/dev/nbd0", 00:07:30.863 "bdev_name": "raid" 00:07:30.863 } 00:07:30.863 ]' 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:30.863 { 00:07:30.863 "nbd_device": "/dev/nbd0", 00:07:30.863 "bdev_name": "raid" 00:07:30.863 } 00:07:30.863 ]' 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:30.863 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:30.864 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:31.123 4096+0 records in 00:07:31.123 4096+0 records out 00:07:31.123 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.032532 s, 64.5 MB/s 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:31.123 4096+0 records in 00:07:31.123 4096+0 records out 00:07:31.123 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.218224 s, 9.6 MB/s 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:31.123 128+0 records in 00:07:31.123 128+0 records out 00:07:31.123 65536 bytes (66 kB, 64 KiB) copied, 0.00111009 s, 59.0 MB/s 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:31.123 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:31.383 2035+0 records in 00:07:31.383 2035+0 records out 00:07:31.383 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.014236 s, 73.2 MB/s 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:31.383 456+0 records in 00:07:31.383 456+0 records out 00:07:31.383 233472 bytes (233 kB, 228 KiB) copied, 0.00347545 s, 67.2 MB/s 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:31.383 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:31.643 [2024-10-29 10:56:36.903382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:31.643 10:56:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:31.643 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:31.643 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:31.643 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 72046 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@952 -- # '[' -z 72046 ']' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # kill -0 72046 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@957 -- # uname 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72046 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:31.902 killing process with pid 72046 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72046' 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@971 -- # kill 72046 00:07:31.902 [2024-10-29 10:56:37.212071] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:31.902 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@976 -- # wait 72046 00:07:31.902 [2024-10-29 10:56:37.212232] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:31.902 [2024-10-29 10:56:37.212301] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:31.902 [2024-10-29 10:56:37.212325] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid, state offline 00:07:31.902 [2024-10-29 10:56:37.253726] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:32.161 ************************************ 00:07:32.161 END TEST raid_function_test_concat 00:07:32.161 10:56:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:07:32.161 00:07:32.161 real 0m2.774s 00:07:32.161 user 0m3.305s 00:07:32.161 sys 0m0.965s 00:07:32.161 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:32.161 10:56:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:32.161 ************************************ 00:07:32.161 10:56:37 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:07:32.161 10:56:37 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:32.161 10:56:37 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:32.161 10:56:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:32.161 ************************************ 00:07:32.161 START TEST raid0_resize_test 00:07:32.161 ************************************ 00:07:32.161 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1127 -- # raid_resize_test 0 00:07:32.161 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:07:32.161 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:32.161 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=72163 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 72163' 00:07:32.162 Process raid pid: 72163 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 72163 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@833 -- # '[' -z 72163 ']' 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:32.162 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:32.162 10:56:37 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.421 [2024-10-29 10:56:37.733205] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:32.421 [2024-10-29 10:56:37.733332] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:32.421 [2024-10-29 10:56:37.905546] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:32.681 [2024-10-29 10:56:37.944693] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:32.681 [2024-10-29 10:56:38.020646] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:32.681 [2024-10-29 10:56:38.020695] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@866 -- # return 0 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.251 Base_1 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.251 Base_2 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.251 [2024-10-29 10:56:38.575827] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:33.251 [2024-10-29 10:56:38.577906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:33.251 [2024-10-29 10:56:38.577965] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:33.251 [2024-10-29 10:56:38.577976] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:33.251 [2024-10-29 10:56:38.578240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:33.251 [2024-10-29 10:56:38.578384] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:33.251 [2024-10-29 10:56:38.578405] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:33.251 [2024-10-29 10:56:38.578541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.251 [2024-10-29 10:56:38.587774] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:33.251 [2024-10-29 10:56:38.587801] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:33.251 true 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.251 [2024-10-29 10:56:38.603922] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:07:33.251 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.252 [2024-10-29 10:56:38.647670] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:33.252 [2024-10-29 10:56:38.647698] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:33.252 [2024-10-29 10:56:38.647726] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:07:33.252 true 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.252 [2024-10-29 10:56:38.663834] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 72163 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@952 -- # '[' -z 72163 ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # kill -0 72163 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@957 -- # uname 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72163 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:33.252 killing process with pid 72163 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72163' 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@971 -- # kill 72163 00:07:33.252 [2024-10-29 10:56:38.747841] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:33.252 [2024-10-29 10:56:38.747951] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:33.252 [2024-10-29 10:56:38.748030] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:33.252 10:56:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@976 -- # wait 72163 00:07:33.252 [2024-10-29 10:56:38.748040] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:33.252 [2024-10-29 10:56:38.750391] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:33.822 10:56:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:33.822 00:07:33.822 real 0m1.422s 00:07:33.822 user 0m1.521s 00:07:33.822 sys 0m0.360s 00:07:33.822 10:56:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:33.822 10:56:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.822 ************************************ 00:07:33.822 END TEST raid0_resize_test 00:07:33.822 ************************************ 00:07:33.822 10:56:39 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:07:33.822 10:56:39 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:07:33.822 10:56:39 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:33.822 10:56:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:33.822 ************************************ 00:07:33.822 START TEST raid1_resize_test 00:07:33.822 ************************************ 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1127 -- # raid_resize_test 1 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=72208 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 72208' 00:07:33.822 Process raid pid: 72208 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 72208 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@833 -- # '[' -z 72208 ']' 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:33.822 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:33.822 10:56:39 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.822 [2024-10-29 10:56:39.222913] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:33.822 [2024-10-29 10:56:39.223051] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:34.081 [2024-10-29 10:56:39.369939] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:34.082 [2024-10-29 10:56:39.407920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.082 [2024-10-29 10:56:39.484424] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.082 [2024-10-29 10:56:39.484474] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@866 -- # return 0 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.658 Base_1 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.658 Base_2 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.658 [2024-10-29 10:56:40.068443] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:34.658 [2024-10-29 10:56:40.071238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:34.658 [2024-10-29 10:56:40.071337] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:34.658 [2024-10-29 10:56:40.071372] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:34.658 [2024-10-29 10:56:40.071793] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005a00 00:07:34.658 [2024-10-29 10:56:40.071986] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:34.658 [2024-10-29 10:56:40.072008] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000006280 00:07:34.658 [2024-10-29 10:56:40.072196] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.658 [2024-10-29 10:56:40.076400] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:34.658 [2024-10-29 10:56:40.076446] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:34.658 true 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.658 [2024-10-29 10:56:40.088660] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:07:34.658 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.659 [2024-10-29 10:56:40.136452] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:34.659 [2024-10-29 10:56:40.136524] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:34.659 [2024-10-29 10:56:40.136592] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:07:34.659 true 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.659 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:34.659 [2024-10-29 10:56:40.148555] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 72208 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@952 -- # '[' -z 72208 ']' 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # kill -0 72208 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@957 -- # uname 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72208 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:34.919 killing process with pid 72208 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72208' 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@971 -- # kill 72208 00:07:34.919 [2024-10-29 10:56:40.234459] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:34.919 [2024-10-29 10:56:40.234612] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:34.919 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@976 -- # wait 72208 00:07:34.919 [2024-10-29 10:56:40.235109] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:34.919 [2024-10-29 10:56:40.235138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Raid, state offline 00:07:34.919 [2024-10-29 10:56:40.236357] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:35.179 10:56:40 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:35.179 00:07:35.179 real 0m1.301s 00:07:35.179 user 0m1.412s 00:07:35.179 sys 0m0.347s 00:07:35.179 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:35.179 10:56:40 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.179 ************************************ 00:07:35.179 END TEST raid1_resize_test 00:07:35.179 ************************************ 00:07:35.179 10:56:40 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:35.179 10:56:40 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:35.179 10:56:40 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:07:35.179 10:56:40 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:35.179 10:56:40 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:35.179 10:56:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:35.179 ************************************ 00:07:35.179 START TEST raid_state_function_test 00:07:35.179 ************************************ 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 2 false 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72260 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72260' 00:07:35.179 Process raid pid: 72260 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72260 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 72260 ']' 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:35.179 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:35.179 10:56:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.179 [2024-10-29 10:56:40.596023] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:35.179 [2024-10-29 10:56:40.596161] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:35.439 [2024-10-29 10:56:40.764870] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:35.439 [2024-10-29 10:56:40.791571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:35.439 [2024-10-29 10:56:40.834863] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:35.439 [2024-10-29 10:56:40.834907] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:36.008 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:36.008 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.009 [2024-10-29 10:56:41.420580] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:36.009 [2024-10-29 10:56:41.420638] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:36.009 [2024-10-29 10:56:41.420654] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:36.009 [2024-10-29 10:56:41.420664] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.009 "name": "Existed_Raid", 00:07:36.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.009 "strip_size_kb": 64, 00:07:36.009 "state": "configuring", 00:07:36.009 "raid_level": "raid0", 00:07:36.009 "superblock": false, 00:07:36.009 "num_base_bdevs": 2, 00:07:36.009 "num_base_bdevs_discovered": 0, 00:07:36.009 "num_base_bdevs_operational": 2, 00:07:36.009 "base_bdevs_list": [ 00:07:36.009 { 00:07:36.009 "name": "BaseBdev1", 00:07:36.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.009 "is_configured": false, 00:07:36.009 "data_offset": 0, 00:07:36.009 "data_size": 0 00:07:36.009 }, 00:07:36.009 { 00:07:36.009 "name": "BaseBdev2", 00:07:36.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.009 "is_configured": false, 00:07:36.009 "data_offset": 0, 00:07:36.009 "data_size": 0 00:07:36.009 } 00:07:36.009 ] 00:07:36.009 }' 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.009 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 [2024-10-29 10:56:41.843832] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:36.581 [2024-10-29 10:56:41.843874] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 [2024-10-29 10:56:41.851771] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:36.581 [2024-10-29 10:56:41.851810] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:36.581 [2024-10-29 10:56:41.851819] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:36.581 [2024-10-29 10:56:41.851828] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 [2024-10-29 10:56:41.868775] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:36.581 BaseBdev1 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 [ 00:07:36.581 { 00:07:36.581 "name": "BaseBdev1", 00:07:36.581 "aliases": [ 00:07:36.581 "4ead2912-450b-4bf9-8cbb-5099495d17be" 00:07:36.581 ], 00:07:36.581 "product_name": "Malloc disk", 00:07:36.581 "block_size": 512, 00:07:36.581 "num_blocks": 65536, 00:07:36.581 "uuid": "4ead2912-450b-4bf9-8cbb-5099495d17be", 00:07:36.581 "assigned_rate_limits": { 00:07:36.581 "rw_ios_per_sec": 0, 00:07:36.581 "rw_mbytes_per_sec": 0, 00:07:36.581 "r_mbytes_per_sec": 0, 00:07:36.581 "w_mbytes_per_sec": 0 00:07:36.581 }, 00:07:36.581 "claimed": true, 00:07:36.581 "claim_type": "exclusive_write", 00:07:36.581 "zoned": false, 00:07:36.581 "supported_io_types": { 00:07:36.581 "read": true, 00:07:36.581 "write": true, 00:07:36.581 "unmap": true, 00:07:36.581 "flush": true, 00:07:36.581 "reset": true, 00:07:36.581 "nvme_admin": false, 00:07:36.581 "nvme_io": false, 00:07:36.581 "nvme_io_md": false, 00:07:36.581 "write_zeroes": true, 00:07:36.581 "zcopy": true, 00:07:36.581 "get_zone_info": false, 00:07:36.581 "zone_management": false, 00:07:36.581 "zone_append": false, 00:07:36.581 "compare": false, 00:07:36.581 "compare_and_write": false, 00:07:36.581 "abort": true, 00:07:36.581 "seek_hole": false, 00:07:36.581 "seek_data": false, 00:07:36.581 "copy": true, 00:07:36.581 "nvme_iov_md": false 00:07:36.581 }, 00:07:36.581 "memory_domains": [ 00:07:36.581 { 00:07:36.581 "dma_device_id": "system", 00:07:36.581 "dma_device_type": 1 00:07:36.581 }, 00:07:36.581 { 00:07:36.581 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.581 "dma_device_type": 2 00:07:36.581 } 00:07:36.581 ], 00:07:36.581 "driver_specific": {} 00:07:36.581 } 00:07:36.581 ] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.581 "name": "Existed_Raid", 00:07:36.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.581 "strip_size_kb": 64, 00:07:36.581 "state": "configuring", 00:07:36.581 "raid_level": "raid0", 00:07:36.581 "superblock": false, 00:07:36.581 "num_base_bdevs": 2, 00:07:36.581 "num_base_bdevs_discovered": 1, 00:07:36.581 "num_base_bdevs_operational": 2, 00:07:36.581 "base_bdevs_list": [ 00:07:36.581 { 00:07:36.581 "name": "BaseBdev1", 00:07:36.581 "uuid": "4ead2912-450b-4bf9-8cbb-5099495d17be", 00:07:36.581 "is_configured": true, 00:07:36.581 "data_offset": 0, 00:07:36.581 "data_size": 65536 00:07:36.581 }, 00:07:36.581 { 00:07:36.581 "name": "BaseBdev2", 00:07:36.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.581 "is_configured": false, 00:07:36.581 "data_offset": 0, 00:07:36.581 "data_size": 0 00:07:36.581 } 00:07:36.581 ] 00:07:36.581 }' 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.581 10:56:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.841 [2024-10-29 10:56:42.336018] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:36.841 [2024-10-29 10:56:42.336080] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.841 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.100 [2024-10-29 10:56:42.344023] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:37.100 [2024-10-29 10:56:42.345910] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:37.101 [2024-10-29 10:56:42.345947] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.101 "name": "Existed_Raid", 00:07:37.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:37.101 "strip_size_kb": 64, 00:07:37.101 "state": "configuring", 00:07:37.101 "raid_level": "raid0", 00:07:37.101 "superblock": false, 00:07:37.101 "num_base_bdevs": 2, 00:07:37.101 "num_base_bdevs_discovered": 1, 00:07:37.101 "num_base_bdevs_operational": 2, 00:07:37.101 "base_bdevs_list": [ 00:07:37.101 { 00:07:37.101 "name": "BaseBdev1", 00:07:37.101 "uuid": "4ead2912-450b-4bf9-8cbb-5099495d17be", 00:07:37.101 "is_configured": true, 00:07:37.101 "data_offset": 0, 00:07:37.101 "data_size": 65536 00:07:37.101 }, 00:07:37.101 { 00:07:37.101 "name": "BaseBdev2", 00:07:37.101 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:37.101 "is_configured": false, 00:07:37.101 "data_offset": 0, 00:07:37.101 "data_size": 0 00:07:37.101 } 00:07:37.101 ] 00:07:37.101 }' 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.101 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.360 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:37.360 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.360 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.360 [2024-10-29 10:56:42.746384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:37.360 [2024-10-29 10:56:42.746430] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:37.360 [2024-10-29 10:56:42.746440] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:37.360 [2024-10-29 10:56:42.746731] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:37.360 [2024-10-29 10:56:42.746919] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:37.360 [2024-10-29 10:56:42.746947] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:37.360 [2024-10-29 10:56:42.747186] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:37.360 BaseBdev2 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.361 [ 00:07:37.361 { 00:07:37.361 "name": "BaseBdev2", 00:07:37.361 "aliases": [ 00:07:37.361 "2f8b965d-cbb8-4b39-ad05-31d81bee6d5b" 00:07:37.361 ], 00:07:37.361 "product_name": "Malloc disk", 00:07:37.361 "block_size": 512, 00:07:37.361 "num_blocks": 65536, 00:07:37.361 "uuid": "2f8b965d-cbb8-4b39-ad05-31d81bee6d5b", 00:07:37.361 "assigned_rate_limits": { 00:07:37.361 "rw_ios_per_sec": 0, 00:07:37.361 "rw_mbytes_per_sec": 0, 00:07:37.361 "r_mbytes_per_sec": 0, 00:07:37.361 "w_mbytes_per_sec": 0 00:07:37.361 }, 00:07:37.361 "claimed": true, 00:07:37.361 "claim_type": "exclusive_write", 00:07:37.361 "zoned": false, 00:07:37.361 "supported_io_types": { 00:07:37.361 "read": true, 00:07:37.361 "write": true, 00:07:37.361 "unmap": true, 00:07:37.361 "flush": true, 00:07:37.361 "reset": true, 00:07:37.361 "nvme_admin": false, 00:07:37.361 "nvme_io": false, 00:07:37.361 "nvme_io_md": false, 00:07:37.361 "write_zeroes": true, 00:07:37.361 "zcopy": true, 00:07:37.361 "get_zone_info": false, 00:07:37.361 "zone_management": false, 00:07:37.361 "zone_append": false, 00:07:37.361 "compare": false, 00:07:37.361 "compare_and_write": false, 00:07:37.361 "abort": true, 00:07:37.361 "seek_hole": false, 00:07:37.361 "seek_data": false, 00:07:37.361 "copy": true, 00:07:37.361 "nvme_iov_md": false 00:07:37.361 }, 00:07:37.361 "memory_domains": [ 00:07:37.361 { 00:07:37.361 "dma_device_id": "system", 00:07:37.361 "dma_device_type": 1 00:07:37.361 }, 00:07:37.361 { 00:07:37.361 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.361 "dma_device_type": 2 00:07:37.361 } 00:07:37.361 ], 00:07:37.361 "driver_specific": {} 00:07:37.361 } 00:07:37.361 ] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.361 "name": "Existed_Raid", 00:07:37.361 "uuid": "fec0281b-054d-4928-8134-6a5290f8a7a1", 00:07:37.361 "strip_size_kb": 64, 00:07:37.361 "state": "online", 00:07:37.361 "raid_level": "raid0", 00:07:37.361 "superblock": false, 00:07:37.361 "num_base_bdevs": 2, 00:07:37.361 "num_base_bdevs_discovered": 2, 00:07:37.361 "num_base_bdevs_operational": 2, 00:07:37.361 "base_bdevs_list": [ 00:07:37.361 { 00:07:37.361 "name": "BaseBdev1", 00:07:37.361 "uuid": "4ead2912-450b-4bf9-8cbb-5099495d17be", 00:07:37.361 "is_configured": true, 00:07:37.361 "data_offset": 0, 00:07:37.361 "data_size": 65536 00:07:37.361 }, 00:07:37.361 { 00:07:37.361 "name": "BaseBdev2", 00:07:37.361 "uuid": "2f8b965d-cbb8-4b39-ad05-31d81bee6d5b", 00:07:37.361 "is_configured": true, 00:07:37.361 "data_offset": 0, 00:07:37.361 "data_size": 65536 00:07:37.361 } 00:07:37.361 ] 00:07:37.361 }' 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.361 10:56:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 [2024-10-29 10:56:43.178021] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:37.930 "name": "Existed_Raid", 00:07:37.930 "aliases": [ 00:07:37.930 "fec0281b-054d-4928-8134-6a5290f8a7a1" 00:07:37.930 ], 00:07:37.930 "product_name": "Raid Volume", 00:07:37.930 "block_size": 512, 00:07:37.930 "num_blocks": 131072, 00:07:37.930 "uuid": "fec0281b-054d-4928-8134-6a5290f8a7a1", 00:07:37.930 "assigned_rate_limits": { 00:07:37.930 "rw_ios_per_sec": 0, 00:07:37.930 "rw_mbytes_per_sec": 0, 00:07:37.930 "r_mbytes_per_sec": 0, 00:07:37.930 "w_mbytes_per_sec": 0 00:07:37.930 }, 00:07:37.930 "claimed": false, 00:07:37.930 "zoned": false, 00:07:37.930 "supported_io_types": { 00:07:37.930 "read": true, 00:07:37.930 "write": true, 00:07:37.930 "unmap": true, 00:07:37.930 "flush": true, 00:07:37.930 "reset": true, 00:07:37.930 "nvme_admin": false, 00:07:37.930 "nvme_io": false, 00:07:37.930 "nvme_io_md": false, 00:07:37.930 "write_zeroes": true, 00:07:37.930 "zcopy": false, 00:07:37.930 "get_zone_info": false, 00:07:37.930 "zone_management": false, 00:07:37.930 "zone_append": false, 00:07:37.930 "compare": false, 00:07:37.930 "compare_and_write": false, 00:07:37.930 "abort": false, 00:07:37.930 "seek_hole": false, 00:07:37.930 "seek_data": false, 00:07:37.930 "copy": false, 00:07:37.930 "nvme_iov_md": false 00:07:37.930 }, 00:07:37.930 "memory_domains": [ 00:07:37.930 { 00:07:37.930 "dma_device_id": "system", 00:07:37.930 "dma_device_type": 1 00:07:37.930 }, 00:07:37.930 { 00:07:37.930 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.930 "dma_device_type": 2 00:07:37.930 }, 00:07:37.930 { 00:07:37.930 "dma_device_id": "system", 00:07:37.930 "dma_device_type": 1 00:07:37.930 }, 00:07:37.930 { 00:07:37.930 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.930 "dma_device_type": 2 00:07:37.930 } 00:07:37.930 ], 00:07:37.930 "driver_specific": { 00:07:37.930 "raid": { 00:07:37.930 "uuid": "fec0281b-054d-4928-8134-6a5290f8a7a1", 00:07:37.930 "strip_size_kb": 64, 00:07:37.930 "state": "online", 00:07:37.930 "raid_level": "raid0", 00:07:37.930 "superblock": false, 00:07:37.930 "num_base_bdevs": 2, 00:07:37.930 "num_base_bdevs_discovered": 2, 00:07:37.930 "num_base_bdevs_operational": 2, 00:07:37.930 "base_bdevs_list": [ 00:07:37.930 { 00:07:37.930 "name": "BaseBdev1", 00:07:37.930 "uuid": "4ead2912-450b-4bf9-8cbb-5099495d17be", 00:07:37.930 "is_configured": true, 00:07:37.930 "data_offset": 0, 00:07:37.930 "data_size": 65536 00:07:37.930 }, 00:07:37.930 { 00:07:37.930 "name": "BaseBdev2", 00:07:37.930 "uuid": "2f8b965d-cbb8-4b39-ad05-31d81bee6d5b", 00:07:37.930 "is_configured": true, 00:07:37.930 "data_offset": 0, 00:07:37.930 "data_size": 65536 00:07:37.930 } 00:07:37.930 ] 00:07:37.930 } 00:07:37.930 } 00:07:37.930 }' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:37.930 BaseBdev2' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 [2024-10-29 10:56:43.401409] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:37.930 [2024-10-29 10:56:43.401447] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:37.930 [2024-10-29 10:56:43.401508] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.930 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:38.252 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.252 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.252 "name": "Existed_Raid", 00:07:38.252 "uuid": "fec0281b-054d-4928-8134-6a5290f8a7a1", 00:07:38.252 "strip_size_kb": 64, 00:07:38.252 "state": "offline", 00:07:38.252 "raid_level": "raid0", 00:07:38.252 "superblock": false, 00:07:38.252 "num_base_bdevs": 2, 00:07:38.252 "num_base_bdevs_discovered": 1, 00:07:38.252 "num_base_bdevs_operational": 1, 00:07:38.252 "base_bdevs_list": [ 00:07:38.252 { 00:07:38.252 "name": null, 00:07:38.252 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.252 "is_configured": false, 00:07:38.252 "data_offset": 0, 00:07:38.252 "data_size": 65536 00:07:38.252 }, 00:07:38.252 { 00:07:38.252 "name": "BaseBdev2", 00:07:38.252 "uuid": "2f8b965d-cbb8-4b39-ad05-31d81bee6d5b", 00:07:38.252 "is_configured": true, 00:07:38.252 "data_offset": 0, 00:07:38.252 "data_size": 65536 00:07:38.252 } 00:07:38.252 ] 00:07:38.252 }' 00:07:38.252 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.252 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.527 [2024-10-29 10:56:43.843950] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:38.527 [2024-10-29 10:56:43.844009] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72260 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 72260 ']' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 72260 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72260 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:38.527 killing process with pid 72260 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72260' 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 72260 00:07:38.527 [2024-10-29 10:56:43.954777] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:38.527 10:56:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 72260 00:07:38.527 [2024-10-29 10:56:43.955852] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:38.787 00:07:38.787 real 0m3.666s 00:07:38.787 user 0m5.799s 00:07:38.787 sys 0m0.712s 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.787 ************************************ 00:07:38.787 END TEST raid_state_function_test 00:07:38.787 ************************************ 00:07:38.787 10:56:44 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:38.787 10:56:44 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:38.787 10:56:44 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:38.787 10:56:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:38.787 ************************************ 00:07:38.787 START TEST raid_state_function_test_sb 00:07:38.787 ************************************ 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 2 true 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72496 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72496' 00:07:38.787 Process raid pid: 72496 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72496 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 72496 ']' 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:38.787 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:38.787 10:56:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.047 [2024-10-29 10:56:44.325642] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:39.047 [2024-10-29 10:56:44.325764] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:39.047 [2024-10-29 10:56:44.497892] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:39.047 [2024-10-29 10:56:44.523138] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:39.306 [2024-10-29 10:56:44.566632] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:39.306 [2024-10-29 10:56:44.566679] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:39.875 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:39.875 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:07:39.875 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.875 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.876 [2024-10-29 10:56:45.192468] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:39.876 [2024-10-29 10:56:45.192526] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:39.876 [2024-10-29 10:56:45.192536] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.876 [2024-10-29 10:56:45.192547] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.876 "name": "Existed_Raid", 00:07:39.876 "uuid": "e2d6da36-04c5-41b9-bfeb-66c9ff8c485e", 00:07:39.876 "strip_size_kb": 64, 00:07:39.876 "state": "configuring", 00:07:39.876 "raid_level": "raid0", 00:07:39.876 "superblock": true, 00:07:39.876 "num_base_bdevs": 2, 00:07:39.876 "num_base_bdevs_discovered": 0, 00:07:39.876 "num_base_bdevs_operational": 2, 00:07:39.876 "base_bdevs_list": [ 00:07:39.876 { 00:07:39.876 "name": "BaseBdev1", 00:07:39.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.876 "is_configured": false, 00:07:39.876 "data_offset": 0, 00:07:39.876 "data_size": 0 00:07:39.876 }, 00:07:39.876 { 00:07:39.876 "name": "BaseBdev2", 00:07:39.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.876 "is_configured": false, 00:07:39.876 "data_offset": 0, 00:07:39.876 "data_size": 0 00:07:39.876 } 00:07:39.876 ] 00:07:39.876 }' 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.876 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.136 [2024-10-29 10:56:45.591668] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:40.136 [2024-10-29 10:56:45.591725] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.136 [2024-10-29 10:56:45.599668] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:40.136 [2024-10-29 10:56:45.599709] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:40.136 [2024-10-29 10:56:45.599717] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:40.136 [2024-10-29 10:56:45.599726] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.136 [2024-10-29 10:56:45.616565] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.136 BaseBdev1 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.136 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.396 [ 00:07:40.396 { 00:07:40.396 "name": "BaseBdev1", 00:07:40.396 "aliases": [ 00:07:40.396 "adcb5428-f012-41cb-b163-4267c61b9dfc" 00:07:40.396 ], 00:07:40.396 "product_name": "Malloc disk", 00:07:40.396 "block_size": 512, 00:07:40.396 "num_blocks": 65536, 00:07:40.396 "uuid": "adcb5428-f012-41cb-b163-4267c61b9dfc", 00:07:40.396 "assigned_rate_limits": { 00:07:40.396 "rw_ios_per_sec": 0, 00:07:40.396 "rw_mbytes_per_sec": 0, 00:07:40.396 "r_mbytes_per_sec": 0, 00:07:40.396 "w_mbytes_per_sec": 0 00:07:40.396 }, 00:07:40.396 "claimed": true, 00:07:40.396 "claim_type": "exclusive_write", 00:07:40.396 "zoned": false, 00:07:40.396 "supported_io_types": { 00:07:40.396 "read": true, 00:07:40.396 "write": true, 00:07:40.396 "unmap": true, 00:07:40.396 "flush": true, 00:07:40.396 "reset": true, 00:07:40.396 "nvme_admin": false, 00:07:40.396 "nvme_io": false, 00:07:40.396 "nvme_io_md": false, 00:07:40.396 "write_zeroes": true, 00:07:40.396 "zcopy": true, 00:07:40.396 "get_zone_info": false, 00:07:40.396 "zone_management": false, 00:07:40.396 "zone_append": false, 00:07:40.396 "compare": false, 00:07:40.396 "compare_and_write": false, 00:07:40.396 "abort": true, 00:07:40.396 "seek_hole": false, 00:07:40.396 "seek_data": false, 00:07:40.396 "copy": true, 00:07:40.396 "nvme_iov_md": false 00:07:40.396 }, 00:07:40.396 "memory_domains": [ 00:07:40.396 { 00:07:40.396 "dma_device_id": "system", 00:07:40.396 "dma_device_type": 1 00:07:40.396 }, 00:07:40.396 { 00:07:40.396 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.396 "dma_device_type": 2 00:07:40.396 } 00:07:40.396 ], 00:07:40.396 "driver_specific": {} 00:07:40.396 } 00:07:40.396 ] 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.396 "name": "Existed_Raid", 00:07:40.396 "uuid": "21c945d7-78e2-4d8d-9fc3-30d56246fe21", 00:07:40.396 "strip_size_kb": 64, 00:07:40.396 "state": "configuring", 00:07:40.396 "raid_level": "raid0", 00:07:40.396 "superblock": true, 00:07:40.396 "num_base_bdevs": 2, 00:07:40.396 "num_base_bdevs_discovered": 1, 00:07:40.396 "num_base_bdevs_operational": 2, 00:07:40.396 "base_bdevs_list": [ 00:07:40.396 { 00:07:40.396 "name": "BaseBdev1", 00:07:40.396 "uuid": "adcb5428-f012-41cb-b163-4267c61b9dfc", 00:07:40.396 "is_configured": true, 00:07:40.396 "data_offset": 2048, 00:07:40.396 "data_size": 63488 00:07:40.396 }, 00:07:40.396 { 00:07:40.396 "name": "BaseBdev2", 00:07:40.396 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.396 "is_configured": false, 00:07:40.396 "data_offset": 0, 00:07:40.396 "data_size": 0 00:07:40.396 } 00:07:40.396 ] 00:07:40.396 }' 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.396 10:56:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.657 [2024-10-29 10:56:46.071906] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:40.657 [2024-10-29 10:56:46.071956] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.657 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.657 [2024-10-29 10:56:46.083910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.657 [2024-10-29 10:56:46.085744] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:40.658 [2024-10-29 10:56:46.085828] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.658 "name": "Existed_Raid", 00:07:40.658 "uuid": "90bba5e1-49b0-4064-9994-2341f17664bb", 00:07:40.658 "strip_size_kb": 64, 00:07:40.658 "state": "configuring", 00:07:40.658 "raid_level": "raid0", 00:07:40.658 "superblock": true, 00:07:40.658 "num_base_bdevs": 2, 00:07:40.658 "num_base_bdevs_discovered": 1, 00:07:40.658 "num_base_bdevs_operational": 2, 00:07:40.658 "base_bdevs_list": [ 00:07:40.658 { 00:07:40.658 "name": "BaseBdev1", 00:07:40.658 "uuid": "adcb5428-f012-41cb-b163-4267c61b9dfc", 00:07:40.658 "is_configured": true, 00:07:40.658 "data_offset": 2048, 00:07:40.658 "data_size": 63488 00:07:40.658 }, 00:07:40.658 { 00:07:40.658 "name": "BaseBdev2", 00:07:40.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.658 "is_configured": false, 00:07:40.658 "data_offset": 0, 00:07:40.658 "data_size": 0 00:07:40.658 } 00:07:40.658 ] 00:07:40.658 }' 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.658 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.229 [2024-10-29 10:56:46.522299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:41.229 [2024-10-29 10:56:46.522607] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:41.229 [2024-10-29 10:56:46.522661] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:41.229 BaseBdev2 00:07:41.229 [2024-10-29 10:56:46.522988] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:41.229 [2024-10-29 10:56:46.523160] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:41.229 [2024-10-29 10:56:46.523220] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:41.229 [2024-10-29 10:56:46.523406] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:41.229 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.230 [ 00:07:41.230 { 00:07:41.230 "name": "BaseBdev2", 00:07:41.230 "aliases": [ 00:07:41.230 "329687e6-99f3-43b3-80e0-17bc8bc0deda" 00:07:41.230 ], 00:07:41.230 "product_name": "Malloc disk", 00:07:41.230 "block_size": 512, 00:07:41.230 "num_blocks": 65536, 00:07:41.230 "uuid": "329687e6-99f3-43b3-80e0-17bc8bc0deda", 00:07:41.230 "assigned_rate_limits": { 00:07:41.230 "rw_ios_per_sec": 0, 00:07:41.230 "rw_mbytes_per_sec": 0, 00:07:41.230 "r_mbytes_per_sec": 0, 00:07:41.230 "w_mbytes_per_sec": 0 00:07:41.230 }, 00:07:41.230 "claimed": true, 00:07:41.230 "claim_type": "exclusive_write", 00:07:41.230 "zoned": false, 00:07:41.230 "supported_io_types": { 00:07:41.230 "read": true, 00:07:41.230 "write": true, 00:07:41.230 "unmap": true, 00:07:41.230 "flush": true, 00:07:41.230 "reset": true, 00:07:41.230 "nvme_admin": false, 00:07:41.230 "nvme_io": false, 00:07:41.230 "nvme_io_md": false, 00:07:41.230 "write_zeroes": true, 00:07:41.230 "zcopy": true, 00:07:41.230 "get_zone_info": false, 00:07:41.230 "zone_management": false, 00:07:41.230 "zone_append": false, 00:07:41.230 "compare": false, 00:07:41.230 "compare_and_write": false, 00:07:41.230 "abort": true, 00:07:41.230 "seek_hole": false, 00:07:41.230 "seek_data": false, 00:07:41.230 "copy": true, 00:07:41.230 "nvme_iov_md": false 00:07:41.230 }, 00:07:41.230 "memory_domains": [ 00:07:41.230 { 00:07:41.230 "dma_device_id": "system", 00:07:41.230 "dma_device_type": 1 00:07:41.230 }, 00:07:41.230 { 00:07:41.230 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.230 "dma_device_type": 2 00:07:41.230 } 00:07:41.230 ], 00:07:41.230 "driver_specific": {} 00:07:41.230 } 00:07:41.230 ] 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:41.230 "name": "Existed_Raid", 00:07:41.230 "uuid": "90bba5e1-49b0-4064-9994-2341f17664bb", 00:07:41.230 "strip_size_kb": 64, 00:07:41.230 "state": "online", 00:07:41.230 "raid_level": "raid0", 00:07:41.230 "superblock": true, 00:07:41.230 "num_base_bdevs": 2, 00:07:41.230 "num_base_bdevs_discovered": 2, 00:07:41.230 "num_base_bdevs_operational": 2, 00:07:41.230 "base_bdevs_list": [ 00:07:41.230 { 00:07:41.230 "name": "BaseBdev1", 00:07:41.230 "uuid": "adcb5428-f012-41cb-b163-4267c61b9dfc", 00:07:41.230 "is_configured": true, 00:07:41.230 "data_offset": 2048, 00:07:41.230 "data_size": 63488 00:07:41.230 }, 00:07:41.230 { 00:07:41.230 "name": "BaseBdev2", 00:07:41.230 "uuid": "329687e6-99f3-43b3-80e0-17bc8bc0deda", 00:07:41.230 "is_configured": true, 00:07:41.230 "data_offset": 2048, 00:07:41.230 "data_size": 63488 00:07:41.230 } 00:07:41.230 ] 00:07:41.230 }' 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:41.230 10:56:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.800 [2024-10-29 10:56:47.013899] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.800 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:41.800 "name": "Existed_Raid", 00:07:41.800 "aliases": [ 00:07:41.800 "90bba5e1-49b0-4064-9994-2341f17664bb" 00:07:41.800 ], 00:07:41.800 "product_name": "Raid Volume", 00:07:41.800 "block_size": 512, 00:07:41.800 "num_blocks": 126976, 00:07:41.800 "uuid": "90bba5e1-49b0-4064-9994-2341f17664bb", 00:07:41.800 "assigned_rate_limits": { 00:07:41.800 "rw_ios_per_sec": 0, 00:07:41.800 "rw_mbytes_per_sec": 0, 00:07:41.800 "r_mbytes_per_sec": 0, 00:07:41.800 "w_mbytes_per_sec": 0 00:07:41.800 }, 00:07:41.800 "claimed": false, 00:07:41.800 "zoned": false, 00:07:41.800 "supported_io_types": { 00:07:41.800 "read": true, 00:07:41.800 "write": true, 00:07:41.800 "unmap": true, 00:07:41.800 "flush": true, 00:07:41.800 "reset": true, 00:07:41.800 "nvme_admin": false, 00:07:41.800 "nvme_io": false, 00:07:41.800 "nvme_io_md": false, 00:07:41.800 "write_zeroes": true, 00:07:41.800 "zcopy": false, 00:07:41.800 "get_zone_info": false, 00:07:41.800 "zone_management": false, 00:07:41.800 "zone_append": false, 00:07:41.800 "compare": false, 00:07:41.800 "compare_and_write": false, 00:07:41.800 "abort": false, 00:07:41.800 "seek_hole": false, 00:07:41.800 "seek_data": false, 00:07:41.800 "copy": false, 00:07:41.800 "nvme_iov_md": false 00:07:41.800 }, 00:07:41.800 "memory_domains": [ 00:07:41.800 { 00:07:41.800 "dma_device_id": "system", 00:07:41.800 "dma_device_type": 1 00:07:41.800 }, 00:07:41.800 { 00:07:41.800 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.800 "dma_device_type": 2 00:07:41.800 }, 00:07:41.800 { 00:07:41.801 "dma_device_id": "system", 00:07:41.801 "dma_device_type": 1 00:07:41.801 }, 00:07:41.801 { 00:07:41.801 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.801 "dma_device_type": 2 00:07:41.801 } 00:07:41.801 ], 00:07:41.801 "driver_specific": { 00:07:41.801 "raid": { 00:07:41.801 "uuid": "90bba5e1-49b0-4064-9994-2341f17664bb", 00:07:41.801 "strip_size_kb": 64, 00:07:41.801 "state": "online", 00:07:41.801 "raid_level": "raid0", 00:07:41.801 "superblock": true, 00:07:41.801 "num_base_bdevs": 2, 00:07:41.801 "num_base_bdevs_discovered": 2, 00:07:41.801 "num_base_bdevs_operational": 2, 00:07:41.801 "base_bdevs_list": [ 00:07:41.801 { 00:07:41.801 "name": "BaseBdev1", 00:07:41.801 "uuid": "adcb5428-f012-41cb-b163-4267c61b9dfc", 00:07:41.801 "is_configured": true, 00:07:41.801 "data_offset": 2048, 00:07:41.801 "data_size": 63488 00:07:41.801 }, 00:07:41.801 { 00:07:41.801 "name": "BaseBdev2", 00:07:41.801 "uuid": "329687e6-99f3-43b3-80e0-17bc8bc0deda", 00:07:41.801 "is_configured": true, 00:07:41.801 "data_offset": 2048, 00:07:41.801 "data_size": 63488 00:07:41.801 } 00:07:41.801 ] 00:07:41.801 } 00:07:41.801 } 00:07:41.801 }' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:41.801 BaseBdev2' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.801 [2024-10-29 10:56:47.233191] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:41.801 [2024-10-29 10:56:47.233288] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:41.801 [2024-10-29 10:56:47.233368] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.801 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.061 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.061 "name": "Existed_Raid", 00:07:42.061 "uuid": "90bba5e1-49b0-4064-9994-2341f17664bb", 00:07:42.061 "strip_size_kb": 64, 00:07:42.061 "state": "offline", 00:07:42.061 "raid_level": "raid0", 00:07:42.061 "superblock": true, 00:07:42.061 "num_base_bdevs": 2, 00:07:42.061 "num_base_bdevs_discovered": 1, 00:07:42.061 "num_base_bdevs_operational": 1, 00:07:42.061 "base_bdevs_list": [ 00:07:42.061 { 00:07:42.061 "name": null, 00:07:42.061 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:42.061 "is_configured": false, 00:07:42.061 "data_offset": 0, 00:07:42.061 "data_size": 63488 00:07:42.061 }, 00:07:42.061 { 00:07:42.061 "name": "BaseBdev2", 00:07:42.061 "uuid": "329687e6-99f3-43b3-80e0-17bc8bc0deda", 00:07:42.061 "is_configured": true, 00:07:42.061 "data_offset": 2048, 00:07:42.061 "data_size": 63488 00:07:42.061 } 00:07:42.061 ] 00:07:42.061 }' 00:07:42.061 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.061 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.321 [2024-10-29 10:56:47.729057] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:42.321 [2024-10-29 10:56:47.729137] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:42.321 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72496 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 72496 ']' 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 72496 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:42.322 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72496 00:07:42.582 killing process with pid 72496 00:07:42.582 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:42.582 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:42.582 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72496' 00:07:42.582 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 72496 00:07:42.582 [2024-10-29 10:56:47.828860] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:42.582 10:56:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 72496 00:07:42.582 [2024-10-29 10:56:47.830480] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:42.842 ************************************ 00:07:42.842 END TEST raid_state_function_test_sb 00:07:42.842 ************************************ 00:07:42.842 10:56:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:42.842 00:07:42.842 real 0m3.919s 00:07:42.842 user 0m6.114s 00:07:42.842 sys 0m0.713s 00:07:42.842 10:56:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:42.842 10:56:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.842 10:56:48 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:42.842 10:56:48 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:07:42.842 10:56:48 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:42.842 10:56:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:42.842 ************************************ 00:07:42.842 START TEST raid_superblock_test 00:07:42.842 ************************************ 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid0 2 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72737 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72737 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 72737 ']' 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:42.842 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:42.842 10:56:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.842 [2024-10-29 10:56:48.305733] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:42.843 [2024-10-29 10:56:48.305955] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72737 ] 00:07:43.102 [2024-10-29 10:56:48.477647] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:43.102 [2024-10-29 10:56:48.517743] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:43.102 [2024-10-29 10:56:48.593984] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.102 [2024-10-29 10:56:48.594122] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.672 malloc1 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.672 [2024-10-29 10:56:49.149573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:43.672 [2024-10-29 10:56:49.149748] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.672 [2024-10-29 10:56:49.149798] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:43.672 [2024-10-29 10:56:49.149842] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.672 [2024-10-29 10:56:49.152378] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.672 [2024-10-29 10:56:49.152465] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:43.672 pt1 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.672 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.932 malloc2 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.932 [2024-10-29 10:56:49.189005] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:43.932 [2024-10-29 10:56:49.189067] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.932 [2024-10-29 10:56:49.189084] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:43.932 [2024-10-29 10:56:49.189095] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.932 [2024-10-29 10:56:49.191539] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.932 [2024-10-29 10:56:49.191638] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:43.932 pt2 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.932 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.932 [2024-10-29 10:56:49.201040] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:43.932 [2024-10-29 10:56:49.203334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:43.933 [2024-10-29 10:56:49.203527] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:07:43.933 [2024-10-29 10:56:49.203587] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:43.933 [2024-10-29 10:56:49.203902] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:43.933 [2024-10-29 10:56:49.204078] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:07:43.933 [2024-10-29 10:56:49.204118] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:07:43.933 [2024-10-29 10:56:49.204369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.933 "name": "raid_bdev1", 00:07:43.933 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:43.933 "strip_size_kb": 64, 00:07:43.933 "state": "online", 00:07:43.933 "raid_level": "raid0", 00:07:43.933 "superblock": true, 00:07:43.933 "num_base_bdevs": 2, 00:07:43.933 "num_base_bdevs_discovered": 2, 00:07:43.933 "num_base_bdevs_operational": 2, 00:07:43.933 "base_bdevs_list": [ 00:07:43.933 { 00:07:43.933 "name": "pt1", 00:07:43.933 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.933 "is_configured": true, 00:07:43.933 "data_offset": 2048, 00:07:43.933 "data_size": 63488 00:07:43.933 }, 00:07:43.933 { 00:07:43.933 "name": "pt2", 00:07:43.933 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.933 "is_configured": true, 00:07:43.933 "data_offset": 2048, 00:07:43.933 "data_size": 63488 00:07:43.933 } 00:07:43.933 ] 00:07:43.933 }' 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.933 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.193 [2024-10-29 10:56:49.656642] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.193 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.452 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:44.452 "name": "raid_bdev1", 00:07:44.452 "aliases": [ 00:07:44.452 "35f4993b-4883-4d7e-aa9a-cef57b3c1098" 00:07:44.452 ], 00:07:44.452 "product_name": "Raid Volume", 00:07:44.452 "block_size": 512, 00:07:44.452 "num_blocks": 126976, 00:07:44.452 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:44.452 "assigned_rate_limits": { 00:07:44.452 "rw_ios_per_sec": 0, 00:07:44.452 "rw_mbytes_per_sec": 0, 00:07:44.452 "r_mbytes_per_sec": 0, 00:07:44.452 "w_mbytes_per_sec": 0 00:07:44.452 }, 00:07:44.452 "claimed": false, 00:07:44.452 "zoned": false, 00:07:44.452 "supported_io_types": { 00:07:44.452 "read": true, 00:07:44.452 "write": true, 00:07:44.452 "unmap": true, 00:07:44.452 "flush": true, 00:07:44.452 "reset": true, 00:07:44.452 "nvme_admin": false, 00:07:44.452 "nvme_io": false, 00:07:44.452 "nvme_io_md": false, 00:07:44.452 "write_zeroes": true, 00:07:44.452 "zcopy": false, 00:07:44.453 "get_zone_info": false, 00:07:44.453 "zone_management": false, 00:07:44.453 "zone_append": false, 00:07:44.453 "compare": false, 00:07:44.453 "compare_and_write": false, 00:07:44.453 "abort": false, 00:07:44.453 "seek_hole": false, 00:07:44.453 "seek_data": false, 00:07:44.453 "copy": false, 00:07:44.453 "nvme_iov_md": false 00:07:44.453 }, 00:07:44.453 "memory_domains": [ 00:07:44.453 { 00:07:44.453 "dma_device_id": "system", 00:07:44.453 "dma_device_type": 1 00:07:44.453 }, 00:07:44.453 { 00:07:44.453 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.453 "dma_device_type": 2 00:07:44.453 }, 00:07:44.453 { 00:07:44.453 "dma_device_id": "system", 00:07:44.453 "dma_device_type": 1 00:07:44.453 }, 00:07:44.453 { 00:07:44.453 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.453 "dma_device_type": 2 00:07:44.453 } 00:07:44.453 ], 00:07:44.453 "driver_specific": { 00:07:44.453 "raid": { 00:07:44.453 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:44.453 "strip_size_kb": 64, 00:07:44.453 "state": "online", 00:07:44.453 "raid_level": "raid0", 00:07:44.453 "superblock": true, 00:07:44.453 "num_base_bdevs": 2, 00:07:44.453 "num_base_bdevs_discovered": 2, 00:07:44.453 "num_base_bdevs_operational": 2, 00:07:44.453 "base_bdevs_list": [ 00:07:44.453 { 00:07:44.453 "name": "pt1", 00:07:44.453 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.453 "is_configured": true, 00:07:44.453 "data_offset": 2048, 00:07:44.453 "data_size": 63488 00:07:44.453 }, 00:07:44.453 { 00:07:44.453 "name": "pt2", 00:07:44.453 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.453 "is_configured": true, 00:07:44.453 "data_offset": 2048, 00:07:44.453 "data_size": 63488 00:07:44.453 } 00:07:44.453 ] 00:07:44.453 } 00:07:44.453 } 00:07:44.453 }' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:44.453 pt2' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.453 [2024-10-29 10:56:49.856152] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=35f4993b-4883-4d7e-aa9a-cef57b3c1098 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 35f4993b-4883-4d7e-aa9a-cef57b3c1098 ']' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.453 [2024-10-29 10:56:49.899827] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.453 [2024-10-29 10:56:49.899935] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:44.453 [2024-10-29 10:56:49.900075] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:44.453 [2024-10-29 10:56:49.900180] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:44.453 [2024-10-29 10:56:49.900249] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.453 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:44.713 10:56:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 [2024-10-29 10:56:50.031622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:44.713 [2024-10-29 10:56:50.034038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:44.713 [2024-10-29 10:56:50.034196] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:44.713 [2024-10-29 10:56:50.034250] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:44.713 [2024-10-29 10:56:50.034268] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.713 [2024-10-29 10:56:50.034284] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:07:44.713 request: 00:07:44.713 { 00:07:44.713 "name": "raid_bdev1", 00:07:44.713 "raid_level": "raid0", 00:07:44.713 "base_bdevs": [ 00:07:44.713 "malloc1", 00:07:44.713 "malloc2" 00:07:44.713 ], 00:07:44.713 "strip_size_kb": 64, 00:07:44.713 "superblock": false, 00:07:44.713 "method": "bdev_raid_create", 00:07:44.713 "req_id": 1 00:07:44.713 } 00:07:44.713 Got JSON-RPC error response 00:07:44.713 response: 00:07:44.713 { 00:07:44.713 "code": -17, 00:07:44.713 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:44.713 } 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 [2024-10-29 10:56:50.091455] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:44.713 [2024-10-29 10:56:50.091506] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.713 [2024-10-29 10:56:50.091528] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:07:44.713 [2024-10-29 10:56:50.091538] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.713 [2024-10-29 10:56:50.094150] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.713 [2024-10-29 10:56:50.094185] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:44.713 [2024-10-29 10:56:50.094263] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:44.713 [2024-10-29 10:56:50.094312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:44.713 pt1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.713 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.713 "name": "raid_bdev1", 00:07:44.713 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:44.713 "strip_size_kb": 64, 00:07:44.713 "state": "configuring", 00:07:44.713 "raid_level": "raid0", 00:07:44.713 "superblock": true, 00:07:44.713 "num_base_bdevs": 2, 00:07:44.713 "num_base_bdevs_discovered": 1, 00:07:44.713 "num_base_bdevs_operational": 2, 00:07:44.713 "base_bdevs_list": [ 00:07:44.713 { 00:07:44.713 "name": "pt1", 00:07:44.713 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.713 "is_configured": true, 00:07:44.713 "data_offset": 2048, 00:07:44.713 "data_size": 63488 00:07:44.713 }, 00:07:44.713 { 00:07:44.713 "name": null, 00:07:44.713 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.713 "is_configured": false, 00:07:44.713 "data_offset": 2048, 00:07:44.714 "data_size": 63488 00:07:44.714 } 00:07:44.714 ] 00:07:44.714 }' 00:07:44.714 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.714 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.282 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.282 [2024-10-29 10:56:50.526801] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:45.282 [2024-10-29 10:56:50.526944] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.282 [2024-10-29 10:56:50.526991] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:07:45.282 [2024-10-29 10:56:50.527020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.282 [2024-10-29 10:56:50.527607] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.282 [2024-10-29 10:56:50.527677] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:45.282 [2024-10-29 10:56:50.527817] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:45.282 [2024-10-29 10:56:50.527882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:45.282 [2024-10-29 10:56:50.528043] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:45.282 [2024-10-29 10:56:50.528080] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:45.282 [2024-10-29 10:56:50.528390] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:07:45.282 [2024-10-29 10:56:50.528559] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:45.282 [2024-10-29 10:56:50.528608] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:45.283 [2024-10-29 10:56:50.528796] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:45.283 pt2 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.283 "name": "raid_bdev1", 00:07:45.283 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:45.283 "strip_size_kb": 64, 00:07:45.283 "state": "online", 00:07:45.283 "raid_level": "raid0", 00:07:45.283 "superblock": true, 00:07:45.283 "num_base_bdevs": 2, 00:07:45.283 "num_base_bdevs_discovered": 2, 00:07:45.283 "num_base_bdevs_operational": 2, 00:07:45.283 "base_bdevs_list": [ 00:07:45.283 { 00:07:45.283 "name": "pt1", 00:07:45.283 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:45.283 "is_configured": true, 00:07:45.283 "data_offset": 2048, 00:07:45.283 "data_size": 63488 00:07:45.283 }, 00:07:45.283 { 00:07:45.283 "name": "pt2", 00:07:45.283 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.283 "is_configured": true, 00:07:45.283 "data_offset": 2048, 00:07:45.283 "data_size": 63488 00:07:45.283 } 00:07:45.283 ] 00:07:45.283 }' 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.283 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.543 [2024-10-29 10:56:50.966330] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:45.543 "name": "raid_bdev1", 00:07:45.543 "aliases": [ 00:07:45.543 "35f4993b-4883-4d7e-aa9a-cef57b3c1098" 00:07:45.543 ], 00:07:45.543 "product_name": "Raid Volume", 00:07:45.543 "block_size": 512, 00:07:45.543 "num_blocks": 126976, 00:07:45.543 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:45.543 "assigned_rate_limits": { 00:07:45.543 "rw_ios_per_sec": 0, 00:07:45.543 "rw_mbytes_per_sec": 0, 00:07:45.543 "r_mbytes_per_sec": 0, 00:07:45.543 "w_mbytes_per_sec": 0 00:07:45.543 }, 00:07:45.543 "claimed": false, 00:07:45.543 "zoned": false, 00:07:45.543 "supported_io_types": { 00:07:45.543 "read": true, 00:07:45.543 "write": true, 00:07:45.543 "unmap": true, 00:07:45.543 "flush": true, 00:07:45.543 "reset": true, 00:07:45.543 "nvme_admin": false, 00:07:45.543 "nvme_io": false, 00:07:45.543 "nvme_io_md": false, 00:07:45.543 "write_zeroes": true, 00:07:45.543 "zcopy": false, 00:07:45.543 "get_zone_info": false, 00:07:45.543 "zone_management": false, 00:07:45.543 "zone_append": false, 00:07:45.543 "compare": false, 00:07:45.543 "compare_and_write": false, 00:07:45.543 "abort": false, 00:07:45.543 "seek_hole": false, 00:07:45.543 "seek_data": false, 00:07:45.543 "copy": false, 00:07:45.543 "nvme_iov_md": false 00:07:45.543 }, 00:07:45.543 "memory_domains": [ 00:07:45.543 { 00:07:45.543 "dma_device_id": "system", 00:07:45.543 "dma_device_type": 1 00:07:45.543 }, 00:07:45.543 { 00:07:45.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.543 "dma_device_type": 2 00:07:45.543 }, 00:07:45.543 { 00:07:45.543 "dma_device_id": "system", 00:07:45.543 "dma_device_type": 1 00:07:45.543 }, 00:07:45.543 { 00:07:45.543 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.543 "dma_device_type": 2 00:07:45.543 } 00:07:45.543 ], 00:07:45.543 "driver_specific": { 00:07:45.543 "raid": { 00:07:45.543 "uuid": "35f4993b-4883-4d7e-aa9a-cef57b3c1098", 00:07:45.543 "strip_size_kb": 64, 00:07:45.543 "state": "online", 00:07:45.543 "raid_level": "raid0", 00:07:45.543 "superblock": true, 00:07:45.543 "num_base_bdevs": 2, 00:07:45.543 "num_base_bdevs_discovered": 2, 00:07:45.543 "num_base_bdevs_operational": 2, 00:07:45.543 "base_bdevs_list": [ 00:07:45.543 { 00:07:45.543 "name": "pt1", 00:07:45.543 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:45.543 "is_configured": true, 00:07:45.543 "data_offset": 2048, 00:07:45.543 "data_size": 63488 00:07:45.543 }, 00:07:45.543 { 00:07:45.543 "name": "pt2", 00:07:45.543 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.543 "is_configured": true, 00:07:45.543 "data_offset": 2048, 00:07:45.543 "data_size": 63488 00:07:45.543 } 00:07:45.543 ] 00:07:45.543 } 00:07:45.543 } 00:07:45.543 }' 00:07:45.543 10:56:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:45.543 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:45.543 pt2' 00:07:45.543 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.803 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:45.803 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.804 [2024-10-29 10:56:51.197910] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 35f4993b-4883-4d7e-aa9a-cef57b3c1098 '!=' 35f4993b-4883-4d7e-aa9a-cef57b3c1098 ']' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72737 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 72737 ']' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 72737 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72737 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:45.804 killing process with pid 72737 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72737' 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 72737 00:07:45.804 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 72737 00:07:45.804 [2024-10-29 10:56:51.274272] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:45.804 [2024-10-29 10:56:51.274450] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.804 [2024-10-29 10:56:51.274532] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.804 [2024-10-29 10:56:51.274542] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:46.064 [2024-10-29 10:56:51.318286] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:46.324 ************************************ 00:07:46.324 END TEST raid_superblock_test 00:07:46.324 ************************************ 00:07:46.324 10:56:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:46.324 00:07:46.324 real 0m3.420s 00:07:46.324 user 0m5.108s 00:07:46.324 sys 0m0.781s 00:07:46.324 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:46.324 10:56:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.324 10:56:51 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:46.324 10:56:51 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:46.324 10:56:51 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:46.324 10:56:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:46.324 ************************************ 00:07:46.324 START TEST raid_read_error_test 00:07:46.324 ************************************ 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 2 read 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.zfyaJSpxyy 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72938 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72938 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 72938 ']' 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:46.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:46.324 10:56:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.324 [2024-10-29 10:56:51.806457] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:46.324 [2024-10-29 10:56:51.806585] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72938 ] 00:07:46.584 [2024-10-29 10:56:51.977343] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:46.584 [2024-10-29 10:56:52.004479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:46.584 [2024-10-29 10:56:52.048438] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.585 [2024-10-29 10:56:52.048476] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.154 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:47.154 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:07:47.154 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.155 BaseBdev1_malloc 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.155 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 true 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 [2024-10-29 10:56:52.667778] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:47.415 [2024-10-29 10:56:52.667838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:47.415 [2024-10-29 10:56:52.667862] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:47.415 [2024-10-29 10:56:52.667873] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:47.415 [2024-10-29 10:56:52.670032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:47.415 [2024-10-29 10:56:52.670075] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:47.415 BaseBdev1 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 BaseBdev2_malloc 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 true 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 [2024-10-29 10:56:52.708560] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:47.415 [2024-10-29 10:56:52.708614] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:47.415 [2024-10-29 10:56:52.708634] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:47.415 [2024-10-29 10:56:52.708644] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:47.415 [2024-10-29 10:56:52.710718] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:47.415 [2024-10-29 10:56:52.710758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:47.415 BaseBdev2 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 [2024-10-29 10:56:52.720589] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:47.415 [2024-10-29 10:56:52.722476] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:47.415 [2024-10-29 10:56:52.722662] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:47.415 [2024-10-29 10:56:52.722676] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:47.415 [2024-10-29 10:56:52.722924] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:47.415 [2024-10-29 10:56:52.723100] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:47.415 [2024-10-29 10:56:52.723115] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:47.415 [2024-10-29 10:56:52.723239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.415 "name": "raid_bdev1", 00:07:47.415 "uuid": "a76ff0ea-5973-49b8-9286-f5b1358a1ef1", 00:07:47.415 "strip_size_kb": 64, 00:07:47.415 "state": "online", 00:07:47.415 "raid_level": "raid0", 00:07:47.415 "superblock": true, 00:07:47.415 "num_base_bdevs": 2, 00:07:47.415 "num_base_bdevs_discovered": 2, 00:07:47.415 "num_base_bdevs_operational": 2, 00:07:47.415 "base_bdevs_list": [ 00:07:47.415 { 00:07:47.415 "name": "BaseBdev1", 00:07:47.415 "uuid": "d995e570-4ec6-56c3-b5dc-3dc19429aff3", 00:07:47.415 "is_configured": true, 00:07:47.415 "data_offset": 2048, 00:07:47.415 "data_size": 63488 00:07:47.415 }, 00:07:47.415 { 00:07:47.415 "name": "BaseBdev2", 00:07:47.415 "uuid": "cf5bfd63-f677-5941-829b-c5c9e3b6fd0c", 00:07:47.415 "is_configured": true, 00:07:47.415 "data_offset": 2048, 00:07:47.415 "data_size": 63488 00:07:47.415 } 00:07:47.415 ] 00:07:47.415 }' 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.415 10:56:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.675 10:56:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:47.935 10:56:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:47.935 [2024-10-29 10:56:53.264139] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.872 "name": "raid_bdev1", 00:07:48.872 "uuid": "a76ff0ea-5973-49b8-9286-f5b1358a1ef1", 00:07:48.872 "strip_size_kb": 64, 00:07:48.872 "state": "online", 00:07:48.872 "raid_level": "raid0", 00:07:48.872 "superblock": true, 00:07:48.872 "num_base_bdevs": 2, 00:07:48.872 "num_base_bdevs_discovered": 2, 00:07:48.872 "num_base_bdevs_operational": 2, 00:07:48.872 "base_bdevs_list": [ 00:07:48.872 { 00:07:48.872 "name": "BaseBdev1", 00:07:48.872 "uuid": "d995e570-4ec6-56c3-b5dc-3dc19429aff3", 00:07:48.872 "is_configured": true, 00:07:48.872 "data_offset": 2048, 00:07:48.872 "data_size": 63488 00:07:48.872 }, 00:07:48.872 { 00:07:48.872 "name": "BaseBdev2", 00:07:48.872 "uuid": "cf5bfd63-f677-5941-829b-c5c9e3b6fd0c", 00:07:48.872 "is_configured": true, 00:07:48.872 "data_offset": 2048, 00:07:48.872 "data_size": 63488 00:07:48.872 } 00:07:48.872 ] 00:07:48.872 }' 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.872 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.131 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.132 [2024-10-29 10:56:54.595592] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:49.132 [2024-10-29 10:56:54.595684] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:49.132 [2024-10-29 10:56:54.598314] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:49.132 [2024-10-29 10:56:54.598421] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:49.132 [2024-10-29 10:56:54.598489] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:49.132 [2024-10-29 10:56:54.598577] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:49.132 { 00:07:49.132 "results": [ 00:07:49.132 { 00:07:49.132 "job": "raid_bdev1", 00:07:49.132 "core_mask": "0x1", 00:07:49.132 "workload": "randrw", 00:07:49.132 "percentage": 50, 00:07:49.132 "status": "finished", 00:07:49.132 "queue_depth": 1, 00:07:49.132 "io_size": 131072, 00:07:49.132 "runtime": 1.332276, 00:07:49.132 "iops": 16773.55142628104, 00:07:49.132 "mibps": 2096.69392828513, 00:07:49.132 "io_failed": 1, 00:07:49.132 "io_timeout": 0, 00:07:49.132 "avg_latency_us": 82.53624579204845, 00:07:49.132 "min_latency_us": 25.2646288209607, 00:07:49.132 "max_latency_us": 1402.2986899563318 00:07:49.132 } 00:07:49.132 ], 00:07:49.132 "core_count": 1 00:07:49.132 } 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72938 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 72938 ']' 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 72938 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:49.132 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 72938 00:07:49.392 killing process with pid 72938 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 72938' 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 72938 00:07:49.392 [2024-10-29 10:56:54.634574] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 72938 00:07:49.392 [2024-10-29 10:56:54.650407] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.zfyaJSpxyy 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:07:49.392 ************************************ 00:07:49.392 END TEST raid_read_error_test 00:07:49.392 ************************************ 00:07:49.392 00:07:49.392 real 0m3.158s 00:07:49.392 user 0m4.011s 00:07:49.392 sys 0m0.483s 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:49.392 10:56:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.652 10:56:54 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:49.652 10:56:54 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:49.652 10:56:54 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:49.652 10:56:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:49.652 ************************************ 00:07:49.652 START TEST raid_write_error_test 00:07:49.652 ************************************ 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 2 write 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.IasLWlizAO 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73067 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73067 00:07:49.652 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 73067 ']' 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:49.652 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:49.653 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:49.653 10:56:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.653 [2024-10-29 10:56:55.038922] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:49.653 [2024-10-29 10:56:55.039472] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73067 ] 00:07:49.912 [2024-10-29 10:56:55.210285] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:49.912 [2024-10-29 10:56:55.239427] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:49.912 [2024-10-29 10:56:55.282533] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:49.912 [2024-10-29 10:56:55.282656] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 BaseBdev1_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 true 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 [2024-10-29 10:56:55.897282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:50.484 [2024-10-29 10:56:55.897334] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:50.484 [2024-10-29 10:56:55.897379] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:50.484 [2024-10-29 10:56:55.897405] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:50.484 [2024-10-29 10:56:55.899498] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:50.484 [2024-10-29 10:56:55.899533] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:50.484 BaseBdev1 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 BaseBdev2_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 true 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 [2024-10-29 10:56:55.937830] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:50.484 [2024-10-29 10:56:55.937878] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:50.484 [2024-10-29 10:56:55.937897] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:50.484 [2024-10-29 10:56:55.937905] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:50.484 [2024-10-29 10:56:55.939958] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:50.484 [2024-10-29 10:56:55.940008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:50.484 BaseBdev2 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 [2024-10-29 10:56:55.949877] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:50.484 [2024-10-29 10:56:55.951764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:50.484 [2024-10-29 10:56:55.951931] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:50.484 [2024-10-29 10:56:55.951944] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:50.484 [2024-10-29 10:56:55.952192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:50.484 [2024-10-29 10:56:55.952322] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:50.484 [2024-10-29 10:56:55.952334] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:07:50.484 [2024-10-29 10:56:55.952493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.484 10:56:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.744 10:56:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.744 "name": "raid_bdev1", 00:07:50.744 "uuid": "d29f5cb6-7457-4d3e-b137-77003ab8283e", 00:07:50.744 "strip_size_kb": 64, 00:07:50.744 "state": "online", 00:07:50.744 "raid_level": "raid0", 00:07:50.744 "superblock": true, 00:07:50.744 "num_base_bdevs": 2, 00:07:50.744 "num_base_bdevs_discovered": 2, 00:07:50.744 "num_base_bdevs_operational": 2, 00:07:50.744 "base_bdevs_list": [ 00:07:50.744 { 00:07:50.744 "name": "BaseBdev1", 00:07:50.744 "uuid": "09bb3bb4-aa27-5083-bb5a-5ca056897f99", 00:07:50.744 "is_configured": true, 00:07:50.744 "data_offset": 2048, 00:07:50.744 "data_size": 63488 00:07:50.744 }, 00:07:50.744 { 00:07:50.744 "name": "BaseBdev2", 00:07:50.744 "uuid": "a39d9467-7266-57ca-8080-a1bf96c324ca", 00:07:50.744 "is_configured": true, 00:07:50.744 "data_offset": 2048, 00:07:50.744 "data_size": 63488 00:07:50.744 } 00:07:50.744 ] 00:07:50.744 }' 00:07:50.744 10:56:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.744 10:56:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.016 10:56:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:51.016 10:56:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:51.016 [2024-10-29 10:56:56.465405] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.987 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.988 "name": "raid_bdev1", 00:07:51.988 "uuid": "d29f5cb6-7457-4d3e-b137-77003ab8283e", 00:07:51.988 "strip_size_kb": 64, 00:07:51.988 "state": "online", 00:07:51.988 "raid_level": "raid0", 00:07:51.988 "superblock": true, 00:07:51.988 "num_base_bdevs": 2, 00:07:51.988 "num_base_bdevs_discovered": 2, 00:07:51.988 "num_base_bdevs_operational": 2, 00:07:51.988 "base_bdevs_list": [ 00:07:51.988 { 00:07:51.988 "name": "BaseBdev1", 00:07:51.988 "uuid": "09bb3bb4-aa27-5083-bb5a-5ca056897f99", 00:07:51.988 "is_configured": true, 00:07:51.988 "data_offset": 2048, 00:07:51.988 "data_size": 63488 00:07:51.988 }, 00:07:51.988 { 00:07:51.988 "name": "BaseBdev2", 00:07:51.988 "uuid": "a39d9467-7266-57ca-8080-a1bf96c324ca", 00:07:51.988 "is_configured": true, 00:07:51.988 "data_offset": 2048, 00:07:51.988 "data_size": 63488 00:07:51.988 } 00:07:51.988 ] 00:07:51.988 }' 00:07:51.988 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.988 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.556 [2024-10-29 10:56:57.851683] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:52.556 [2024-10-29 10:56:57.851783] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:52.556 [2024-10-29 10:56:57.854299] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:52.556 [2024-10-29 10:56:57.854408] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:52.556 [2024-10-29 10:56:57.854463] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:52.556 [2024-10-29 10:56:57.854518] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:07:52.556 { 00:07:52.556 "results": [ 00:07:52.556 { 00:07:52.556 "job": "raid_bdev1", 00:07:52.556 "core_mask": "0x1", 00:07:52.556 "workload": "randrw", 00:07:52.556 "percentage": 50, 00:07:52.556 "status": "finished", 00:07:52.556 "queue_depth": 1, 00:07:52.556 "io_size": 131072, 00:07:52.556 "runtime": 1.387124, 00:07:52.556 "iops": 17265.94017549981, 00:07:52.556 "mibps": 2158.2425219374763, 00:07:52.556 "io_failed": 1, 00:07:52.556 "io_timeout": 0, 00:07:52.556 "avg_latency_us": 80.27610199061803, 00:07:52.556 "min_latency_us": 24.258515283842794, 00:07:52.556 "max_latency_us": 1345.0620087336245 00:07:52.556 } 00:07:52.556 ], 00:07:52.556 "core_count": 1 00:07:52.556 } 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73067 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 73067 ']' 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 73067 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73067 00:07:52.556 killing process with pid 73067 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73067' 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 73067 00:07:52.556 [2024-10-29 10:56:57.888016] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:52.556 10:56:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 73067 00:07:52.556 [2024-10-29 10:56:57.903870] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.IasLWlizAO 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:52.816 ************************************ 00:07:52.816 END TEST raid_write_error_test 00:07:52.816 ************************************ 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:52.816 00:07:52.816 real 0m3.182s 00:07:52.816 user 0m4.050s 00:07:52.816 sys 0m0.510s 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:52.816 10:56:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.816 10:56:58 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:52.816 10:56:58 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:52.816 10:56:58 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:52.816 10:56:58 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:52.816 10:56:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:52.816 ************************************ 00:07:52.816 START TEST raid_state_function_test 00:07:52.816 ************************************ 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 2 false 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:52.816 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73194 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73194' 00:07:52.817 Process raid pid: 73194 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73194 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 73194 ']' 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:52.817 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:52.817 10:56:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.817 [2024-10-29 10:56:58.281234] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:52.817 [2024-10-29 10:56:58.281387] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:53.076 [2024-10-29 10:56:58.431459] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:53.076 [2024-10-29 10:56:58.456051] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:53.076 [2024-10-29 10:56:58.498629] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:53.076 [2024-10-29 10:56:58.498663] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.645 [2024-10-29 10:56:59.104527] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:53.645 [2024-10-29 10:56:59.104643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:53.645 [2024-10-29 10:56:59.104664] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:53.645 [2024-10-29 10:56:59.104675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.645 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.646 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.646 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.905 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.905 "name": "Existed_Raid", 00:07:53.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.905 "strip_size_kb": 64, 00:07:53.905 "state": "configuring", 00:07:53.905 "raid_level": "concat", 00:07:53.905 "superblock": false, 00:07:53.905 "num_base_bdevs": 2, 00:07:53.905 "num_base_bdevs_discovered": 0, 00:07:53.905 "num_base_bdevs_operational": 2, 00:07:53.905 "base_bdevs_list": [ 00:07:53.905 { 00:07:53.905 "name": "BaseBdev1", 00:07:53.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.905 "is_configured": false, 00:07:53.905 "data_offset": 0, 00:07:53.905 "data_size": 0 00:07:53.905 }, 00:07:53.905 { 00:07:53.905 "name": "BaseBdev2", 00:07:53.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.905 "is_configured": false, 00:07:53.905 "data_offset": 0, 00:07:53.905 "data_size": 0 00:07:53.905 } 00:07:53.905 ] 00:07:53.905 }' 00:07:53.905 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.905 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.164 [2024-10-29 10:56:59.551665] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:54.164 [2024-10-29 10:56:59.551776] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.164 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.164 [2024-10-29 10:56:59.559653] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:54.164 [2024-10-29 10:56:59.559733] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:54.164 [2024-10-29 10:56:59.559759] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:54.165 [2024-10-29 10:56:59.559781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.165 [2024-10-29 10:56:59.576613] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:54.165 BaseBdev1 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.165 [ 00:07:54.165 { 00:07:54.165 "name": "BaseBdev1", 00:07:54.165 "aliases": [ 00:07:54.165 "7b9d5388-d1d2-43bc-b659-158f3a38b271" 00:07:54.165 ], 00:07:54.165 "product_name": "Malloc disk", 00:07:54.165 "block_size": 512, 00:07:54.165 "num_blocks": 65536, 00:07:54.165 "uuid": "7b9d5388-d1d2-43bc-b659-158f3a38b271", 00:07:54.165 "assigned_rate_limits": { 00:07:54.165 "rw_ios_per_sec": 0, 00:07:54.165 "rw_mbytes_per_sec": 0, 00:07:54.165 "r_mbytes_per_sec": 0, 00:07:54.165 "w_mbytes_per_sec": 0 00:07:54.165 }, 00:07:54.165 "claimed": true, 00:07:54.165 "claim_type": "exclusive_write", 00:07:54.165 "zoned": false, 00:07:54.165 "supported_io_types": { 00:07:54.165 "read": true, 00:07:54.165 "write": true, 00:07:54.165 "unmap": true, 00:07:54.165 "flush": true, 00:07:54.165 "reset": true, 00:07:54.165 "nvme_admin": false, 00:07:54.165 "nvme_io": false, 00:07:54.165 "nvme_io_md": false, 00:07:54.165 "write_zeroes": true, 00:07:54.165 "zcopy": true, 00:07:54.165 "get_zone_info": false, 00:07:54.165 "zone_management": false, 00:07:54.165 "zone_append": false, 00:07:54.165 "compare": false, 00:07:54.165 "compare_and_write": false, 00:07:54.165 "abort": true, 00:07:54.165 "seek_hole": false, 00:07:54.165 "seek_data": false, 00:07:54.165 "copy": true, 00:07:54.165 "nvme_iov_md": false 00:07:54.165 }, 00:07:54.165 "memory_domains": [ 00:07:54.165 { 00:07:54.165 "dma_device_id": "system", 00:07:54.165 "dma_device_type": 1 00:07:54.165 }, 00:07:54.165 { 00:07:54.165 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.165 "dma_device_type": 2 00:07:54.165 } 00:07:54.165 ], 00:07:54.165 "driver_specific": {} 00:07:54.165 } 00:07:54.165 ] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.165 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.423 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.423 "name": "Existed_Raid", 00:07:54.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.423 "strip_size_kb": 64, 00:07:54.423 "state": "configuring", 00:07:54.423 "raid_level": "concat", 00:07:54.423 "superblock": false, 00:07:54.423 "num_base_bdevs": 2, 00:07:54.423 "num_base_bdevs_discovered": 1, 00:07:54.423 "num_base_bdevs_operational": 2, 00:07:54.423 "base_bdevs_list": [ 00:07:54.423 { 00:07:54.423 "name": "BaseBdev1", 00:07:54.423 "uuid": "7b9d5388-d1d2-43bc-b659-158f3a38b271", 00:07:54.423 "is_configured": true, 00:07:54.423 "data_offset": 0, 00:07:54.423 "data_size": 65536 00:07:54.423 }, 00:07:54.423 { 00:07:54.423 "name": "BaseBdev2", 00:07:54.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.423 "is_configured": false, 00:07:54.423 "data_offset": 0, 00:07:54.423 "data_size": 0 00:07:54.423 } 00:07:54.423 ] 00:07:54.423 }' 00:07:54.423 10:56:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.423 10:56:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.683 [2024-10-29 10:57:00.007962] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:54.683 [2024-10-29 10:57:00.008024] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.683 [2024-10-29 10:57:00.019947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:54.683 [2024-10-29 10:57:00.021760] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:54.683 [2024-10-29 10:57:00.021801] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:54.683 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.684 "name": "Existed_Raid", 00:07:54.684 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.684 "strip_size_kb": 64, 00:07:54.684 "state": "configuring", 00:07:54.684 "raid_level": "concat", 00:07:54.684 "superblock": false, 00:07:54.684 "num_base_bdevs": 2, 00:07:54.684 "num_base_bdevs_discovered": 1, 00:07:54.684 "num_base_bdevs_operational": 2, 00:07:54.684 "base_bdevs_list": [ 00:07:54.684 { 00:07:54.684 "name": "BaseBdev1", 00:07:54.684 "uuid": "7b9d5388-d1d2-43bc-b659-158f3a38b271", 00:07:54.684 "is_configured": true, 00:07:54.684 "data_offset": 0, 00:07:54.684 "data_size": 65536 00:07:54.684 }, 00:07:54.684 { 00:07:54.684 "name": "BaseBdev2", 00:07:54.684 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.684 "is_configured": false, 00:07:54.684 "data_offset": 0, 00:07:54.684 "data_size": 0 00:07:54.684 } 00:07:54.684 ] 00:07:54.684 }' 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.684 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.944 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:54.944 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.944 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.204 [2024-10-29 10:57:00.454324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:55.205 [2024-10-29 10:57:00.454469] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:55.205 [2024-10-29 10:57:00.454508] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:55.205 [2024-10-29 10:57:00.454859] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:55.205 [2024-10-29 10:57:00.455068] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:55.205 [2024-10-29 10:57:00.455133] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:55.205 [2024-10-29 10:57:00.455438] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:55.205 BaseBdev2 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.205 [ 00:07:55.205 { 00:07:55.205 "name": "BaseBdev2", 00:07:55.205 "aliases": [ 00:07:55.205 "a0a1acd4-410f-4d67-bea1-7a4cae9eb866" 00:07:55.205 ], 00:07:55.205 "product_name": "Malloc disk", 00:07:55.205 "block_size": 512, 00:07:55.205 "num_blocks": 65536, 00:07:55.205 "uuid": "a0a1acd4-410f-4d67-bea1-7a4cae9eb866", 00:07:55.205 "assigned_rate_limits": { 00:07:55.205 "rw_ios_per_sec": 0, 00:07:55.205 "rw_mbytes_per_sec": 0, 00:07:55.205 "r_mbytes_per_sec": 0, 00:07:55.205 "w_mbytes_per_sec": 0 00:07:55.205 }, 00:07:55.205 "claimed": true, 00:07:55.205 "claim_type": "exclusive_write", 00:07:55.205 "zoned": false, 00:07:55.205 "supported_io_types": { 00:07:55.205 "read": true, 00:07:55.205 "write": true, 00:07:55.205 "unmap": true, 00:07:55.205 "flush": true, 00:07:55.205 "reset": true, 00:07:55.205 "nvme_admin": false, 00:07:55.205 "nvme_io": false, 00:07:55.205 "nvme_io_md": false, 00:07:55.205 "write_zeroes": true, 00:07:55.205 "zcopy": true, 00:07:55.205 "get_zone_info": false, 00:07:55.205 "zone_management": false, 00:07:55.205 "zone_append": false, 00:07:55.205 "compare": false, 00:07:55.205 "compare_and_write": false, 00:07:55.205 "abort": true, 00:07:55.205 "seek_hole": false, 00:07:55.205 "seek_data": false, 00:07:55.205 "copy": true, 00:07:55.205 "nvme_iov_md": false 00:07:55.205 }, 00:07:55.205 "memory_domains": [ 00:07:55.205 { 00:07:55.205 "dma_device_id": "system", 00:07:55.205 "dma_device_type": 1 00:07:55.205 }, 00:07:55.205 { 00:07:55.205 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.205 "dma_device_type": 2 00:07:55.205 } 00:07:55.205 ], 00:07:55.205 "driver_specific": {} 00:07:55.205 } 00:07:55.205 ] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.205 "name": "Existed_Raid", 00:07:55.205 "uuid": "45e9c051-9605-4be8-af6f-22f62e14e596", 00:07:55.205 "strip_size_kb": 64, 00:07:55.205 "state": "online", 00:07:55.205 "raid_level": "concat", 00:07:55.205 "superblock": false, 00:07:55.205 "num_base_bdevs": 2, 00:07:55.205 "num_base_bdevs_discovered": 2, 00:07:55.205 "num_base_bdevs_operational": 2, 00:07:55.205 "base_bdevs_list": [ 00:07:55.205 { 00:07:55.205 "name": "BaseBdev1", 00:07:55.205 "uuid": "7b9d5388-d1d2-43bc-b659-158f3a38b271", 00:07:55.205 "is_configured": true, 00:07:55.205 "data_offset": 0, 00:07:55.205 "data_size": 65536 00:07:55.205 }, 00:07:55.205 { 00:07:55.205 "name": "BaseBdev2", 00:07:55.205 "uuid": "a0a1acd4-410f-4d67-bea1-7a4cae9eb866", 00:07:55.205 "is_configured": true, 00:07:55.205 "data_offset": 0, 00:07:55.205 "data_size": 65536 00:07:55.205 } 00:07:55.205 ] 00:07:55.205 }' 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.205 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.465 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.726 [2024-10-29 10:57:00.965785] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:55.726 10:57:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.726 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:55.726 "name": "Existed_Raid", 00:07:55.726 "aliases": [ 00:07:55.726 "45e9c051-9605-4be8-af6f-22f62e14e596" 00:07:55.726 ], 00:07:55.726 "product_name": "Raid Volume", 00:07:55.726 "block_size": 512, 00:07:55.726 "num_blocks": 131072, 00:07:55.726 "uuid": "45e9c051-9605-4be8-af6f-22f62e14e596", 00:07:55.726 "assigned_rate_limits": { 00:07:55.726 "rw_ios_per_sec": 0, 00:07:55.726 "rw_mbytes_per_sec": 0, 00:07:55.726 "r_mbytes_per_sec": 0, 00:07:55.726 "w_mbytes_per_sec": 0 00:07:55.726 }, 00:07:55.726 "claimed": false, 00:07:55.726 "zoned": false, 00:07:55.726 "supported_io_types": { 00:07:55.726 "read": true, 00:07:55.726 "write": true, 00:07:55.726 "unmap": true, 00:07:55.726 "flush": true, 00:07:55.726 "reset": true, 00:07:55.726 "nvme_admin": false, 00:07:55.726 "nvme_io": false, 00:07:55.726 "nvme_io_md": false, 00:07:55.726 "write_zeroes": true, 00:07:55.726 "zcopy": false, 00:07:55.726 "get_zone_info": false, 00:07:55.726 "zone_management": false, 00:07:55.726 "zone_append": false, 00:07:55.726 "compare": false, 00:07:55.726 "compare_and_write": false, 00:07:55.726 "abort": false, 00:07:55.726 "seek_hole": false, 00:07:55.726 "seek_data": false, 00:07:55.726 "copy": false, 00:07:55.726 "nvme_iov_md": false 00:07:55.726 }, 00:07:55.726 "memory_domains": [ 00:07:55.726 { 00:07:55.726 "dma_device_id": "system", 00:07:55.726 "dma_device_type": 1 00:07:55.726 }, 00:07:55.726 { 00:07:55.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.726 "dma_device_type": 2 00:07:55.726 }, 00:07:55.726 { 00:07:55.726 "dma_device_id": "system", 00:07:55.726 "dma_device_type": 1 00:07:55.726 }, 00:07:55.726 { 00:07:55.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.726 "dma_device_type": 2 00:07:55.726 } 00:07:55.726 ], 00:07:55.726 "driver_specific": { 00:07:55.726 "raid": { 00:07:55.726 "uuid": "45e9c051-9605-4be8-af6f-22f62e14e596", 00:07:55.726 "strip_size_kb": 64, 00:07:55.726 "state": "online", 00:07:55.726 "raid_level": "concat", 00:07:55.726 "superblock": false, 00:07:55.726 "num_base_bdevs": 2, 00:07:55.726 "num_base_bdevs_discovered": 2, 00:07:55.726 "num_base_bdevs_operational": 2, 00:07:55.726 "base_bdevs_list": [ 00:07:55.726 { 00:07:55.726 "name": "BaseBdev1", 00:07:55.726 "uuid": "7b9d5388-d1d2-43bc-b659-158f3a38b271", 00:07:55.726 "is_configured": true, 00:07:55.726 "data_offset": 0, 00:07:55.726 "data_size": 65536 00:07:55.726 }, 00:07:55.726 { 00:07:55.726 "name": "BaseBdev2", 00:07:55.726 "uuid": "a0a1acd4-410f-4d67-bea1-7a4cae9eb866", 00:07:55.726 "is_configured": true, 00:07:55.726 "data_offset": 0, 00:07:55.726 "data_size": 65536 00:07:55.726 } 00:07:55.726 ] 00:07:55.726 } 00:07:55.726 } 00:07:55.726 }' 00:07:55.726 10:57:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:55.726 BaseBdev2' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:55.726 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.727 [2024-10-29 10:57:01.169188] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:55.727 [2024-10-29 10:57:01.169224] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:55.727 [2024-10-29 10:57:01.169278] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.727 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.987 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.987 "name": "Existed_Raid", 00:07:55.987 "uuid": "45e9c051-9605-4be8-af6f-22f62e14e596", 00:07:55.987 "strip_size_kb": 64, 00:07:55.987 "state": "offline", 00:07:55.987 "raid_level": "concat", 00:07:55.987 "superblock": false, 00:07:55.987 "num_base_bdevs": 2, 00:07:55.987 "num_base_bdevs_discovered": 1, 00:07:55.987 "num_base_bdevs_operational": 1, 00:07:55.987 "base_bdevs_list": [ 00:07:55.987 { 00:07:55.987 "name": null, 00:07:55.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.987 "is_configured": false, 00:07:55.987 "data_offset": 0, 00:07:55.987 "data_size": 65536 00:07:55.987 }, 00:07:55.987 { 00:07:55.987 "name": "BaseBdev2", 00:07:55.987 "uuid": "a0a1acd4-410f-4d67-bea1-7a4cae9eb866", 00:07:55.987 "is_configured": true, 00:07:55.987 "data_offset": 0, 00:07:55.987 "data_size": 65536 00:07:55.987 } 00:07:55.987 ] 00:07:55.987 }' 00:07:55.987 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.987 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.248 [2024-10-29 10:57:01.675451] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:56.248 [2024-10-29 10:57:01.675555] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73194 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 73194 ']' 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 73194 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:07:56.248 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73194 00:07:56.508 killing process with pid 73194 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73194' 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 73194 00:07:56.508 [2024-10-29 10:57:01.780750] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 73194 00:07:56.508 [2024-10-29 10:57:01.781738] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:56.508 00:07:56.508 real 0m3.810s 00:07:56.508 user 0m6.029s 00:07:56.508 sys 0m0.740s 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:07:56.508 10:57:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.508 ************************************ 00:07:56.508 END TEST raid_state_function_test 00:07:56.508 ************************************ 00:07:56.769 10:57:02 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:56.769 10:57:02 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:07:56.769 10:57:02 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:07:56.769 10:57:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:56.769 ************************************ 00:07:56.769 START TEST raid_state_function_test_sb 00:07:56.769 ************************************ 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 2 true 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73436 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73436' 00:07:56.769 Process raid pid: 73436 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73436 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 73436 ']' 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:56.769 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:07:56.769 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.769 [2024-10-29 10:57:02.156165] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:07:56.769 [2024-10-29 10:57:02.156390] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:57.029 [2024-10-29 10:57:02.327512] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:57.029 [2024-10-29 10:57:02.356622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:57.029 [2024-10-29 10:57:02.399451] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:57.029 [2024-10-29 10:57:02.399572] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.599 [2024-10-29 10:57:02.988998] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:57.599 [2024-10-29 10:57:02.989113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:57.599 [2024-10-29 10:57:02.989158] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:57.599 [2024-10-29 10:57:02.989186] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.599 10:57:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.599 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.599 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.599 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.599 "name": "Existed_Raid", 00:07:57.599 "uuid": "2e0201f8-966b-4862-907a-6c05ef2045f6", 00:07:57.599 "strip_size_kb": 64, 00:07:57.599 "state": "configuring", 00:07:57.599 "raid_level": "concat", 00:07:57.599 "superblock": true, 00:07:57.599 "num_base_bdevs": 2, 00:07:57.599 "num_base_bdevs_discovered": 0, 00:07:57.599 "num_base_bdevs_operational": 2, 00:07:57.599 "base_bdevs_list": [ 00:07:57.599 { 00:07:57.599 "name": "BaseBdev1", 00:07:57.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.599 "is_configured": false, 00:07:57.599 "data_offset": 0, 00:07:57.599 "data_size": 0 00:07:57.599 }, 00:07:57.599 { 00:07:57.599 "name": "BaseBdev2", 00:07:57.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.599 "is_configured": false, 00:07:57.599 "data_offset": 0, 00:07:57.599 "data_size": 0 00:07:57.599 } 00:07:57.599 ] 00:07:57.599 }' 00:07:57.599 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.599 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 [2024-10-29 10:57:03.436115] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:58.170 [2024-10-29 10:57:03.436163] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 [2024-10-29 10:57:03.448099] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:58.170 [2024-10-29 10:57:03.448143] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:58.170 [2024-10-29 10:57:03.448152] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:58.170 [2024-10-29 10:57:03.448161] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 [2024-10-29 10:57:03.469340] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:58.170 BaseBdev1 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.170 [ 00:07:58.170 { 00:07:58.170 "name": "BaseBdev1", 00:07:58.170 "aliases": [ 00:07:58.170 "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29" 00:07:58.170 ], 00:07:58.170 "product_name": "Malloc disk", 00:07:58.170 "block_size": 512, 00:07:58.170 "num_blocks": 65536, 00:07:58.170 "uuid": "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29", 00:07:58.170 "assigned_rate_limits": { 00:07:58.170 "rw_ios_per_sec": 0, 00:07:58.170 "rw_mbytes_per_sec": 0, 00:07:58.170 "r_mbytes_per_sec": 0, 00:07:58.170 "w_mbytes_per_sec": 0 00:07:58.170 }, 00:07:58.170 "claimed": true, 00:07:58.170 "claim_type": "exclusive_write", 00:07:58.170 "zoned": false, 00:07:58.170 "supported_io_types": { 00:07:58.170 "read": true, 00:07:58.170 "write": true, 00:07:58.170 "unmap": true, 00:07:58.170 "flush": true, 00:07:58.170 "reset": true, 00:07:58.170 "nvme_admin": false, 00:07:58.170 "nvme_io": false, 00:07:58.170 "nvme_io_md": false, 00:07:58.170 "write_zeroes": true, 00:07:58.170 "zcopy": true, 00:07:58.170 "get_zone_info": false, 00:07:58.170 "zone_management": false, 00:07:58.170 "zone_append": false, 00:07:58.170 "compare": false, 00:07:58.170 "compare_and_write": false, 00:07:58.170 "abort": true, 00:07:58.170 "seek_hole": false, 00:07:58.170 "seek_data": false, 00:07:58.170 "copy": true, 00:07:58.170 "nvme_iov_md": false 00:07:58.170 }, 00:07:58.170 "memory_domains": [ 00:07:58.170 { 00:07:58.170 "dma_device_id": "system", 00:07:58.170 "dma_device_type": 1 00:07:58.170 }, 00:07:58.170 { 00:07:58.170 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.170 "dma_device_type": 2 00:07:58.170 } 00:07:58.170 ], 00:07:58.170 "driver_specific": {} 00:07:58.170 } 00:07:58.170 ] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.170 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.171 "name": "Existed_Raid", 00:07:58.171 "uuid": "e00042df-ee67-4b8f-a7a0-fa7a7efed559", 00:07:58.171 "strip_size_kb": 64, 00:07:58.171 "state": "configuring", 00:07:58.171 "raid_level": "concat", 00:07:58.171 "superblock": true, 00:07:58.171 "num_base_bdevs": 2, 00:07:58.171 "num_base_bdevs_discovered": 1, 00:07:58.171 "num_base_bdevs_operational": 2, 00:07:58.171 "base_bdevs_list": [ 00:07:58.171 { 00:07:58.171 "name": "BaseBdev1", 00:07:58.171 "uuid": "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29", 00:07:58.171 "is_configured": true, 00:07:58.171 "data_offset": 2048, 00:07:58.171 "data_size": 63488 00:07:58.171 }, 00:07:58.171 { 00:07:58.171 "name": "BaseBdev2", 00:07:58.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.171 "is_configured": false, 00:07:58.171 "data_offset": 0, 00:07:58.171 "data_size": 0 00:07:58.171 } 00:07:58.171 ] 00:07:58.171 }' 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.171 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.741 [2024-10-29 10:57:03.936591] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:58.741 [2024-10-29 10:57:03.936718] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.741 [2024-10-29 10:57:03.948627] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:58.741 [2024-10-29 10:57:03.950491] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:58.741 [2024-10-29 10:57:03.950538] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.741 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.742 10:57:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.742 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.742 "name": "Existed_Raid", 00:07:58.742 "uuid": "a25aa725-e1ea-46fb-a9c3-b8af412913d6", 00:07:58.742 "strip_size_kb": 64, 00:07:58.742 "state": "configuring", 00:07:58.742 "raid_level": "concat", 00:07:58.742 "superblock": true, 00:07:58.742 "num_base_bdevs": 2, 00:07:58.742 "num_base_bdevs_discovered": 1, 00:07:58.742 "num_base_bdevs_operational": 2, 00:07:58.742 "base_bdevs_list": [ 00:07:58.742 { 00:07:58.742 "name": "BaseBdev1", 00:07:58.742 "uuid": "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29", 00:07:58.742 "is_configured": true, 00:07:58.742 "data_offset": 2048, 00:07:58.742 "data_size": 63488 00:07:58.742 }, 00:07:58.742 { 00:07:58.742 "name": "BaseBdev2", 00:07:58.742 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.742 "is_configured": false, 00:07:58.742 "data_offset": 0, 00:07:58.742 "data_size": 0 00:07:58.742 } 00:07:58.742 ] 00:07:58.742 }' 00:07:58.742 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.742 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.003 [2024-10-29 10:57:04.370978] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:59.003 [2024-10-29 10:57:04.371307] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:07:59.003 [2024-10-29 10:57:04.371374] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:59.003 BaseBdev2 00:07:59.003 [2024-10-29 10:57:04.371718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:07:59.003 [2024-10-29 10:57:04.371914] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:07:59.003 [2024-10-29 10:57:04.371965] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:07:59.003 [2024-10-29 10:57:04.372139] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.003 [ 00:07:59.003 { 00:07:59.003 "name": "BaseBdev2", 00:07:59.003 "aliases": [ 00:07:59.003 "15868bd1-02c6-4997-a975-6df08edded1d" 00:07:59.003 ], 00:07:59.003 "product_name": "Malloc disk", 00:07:59.003 "block_size": 512, 00:07:59.003 "num_blocks": 65536, 00:07:59.003 "uuid": "15868bd1-02c6-4997-a975-6df08edded1d", 00:07:59.003 "assigned_rate_limits": { 00:07:59.003 "rw_ios_per_sec": 0, 00:07:59.003 "rw_mbytes_per_sec": 0, 00:07:59.003 "r_mbytes_per_sec": 0, 00:07:59.003 "w_mbytes_per_sec": 0 00:07:59.003 }, 00:07:59.003 "claimed": true, 00:07:59.003 "claim_type": "exclusive_write", 00:07:59.003 "zoned": false, 00:07:59.003 "supported_io_types": { 00:07:59.003 "read": true, 00:07:59.003 "write": true, 00:07:59.003 "unmap": true, 00:07:59.003 "flush": true, 00:07:59.003 "reset": true, 00:07:59.003 "nvme_admin": false, 00:07:59.003 "nvme_io": false, 00:07:59.003 "nvme_io_md": false, 00:07:59.003 "write_zeroes": true, 00:07:59.003 "zcopy": true, 00:07:59.003 "get_zone_info": false, 00:07:59.003 "zone_management": false, 00:07:59.003 "zone_append": false, 00:07:59.003 "compare": false, 00:07:59.003 "compare_and_write": false, 00:07:59.003 "abort": true, 00:07:59.003 "seek_hole": false, 00:07:59.003 "seek_data": false, 00:07:59.003 "copy": true, 00:07:59.003 "nvme_iov_md": false 00:07:59.003 }, 00:07:59.003 "memory_domains": [ 00:07:59.003 { 00:07:59.003 "dma_device_id": "system", 00:07:59.003 "dma_device_type": 1 00:07:59.003 }, 00:07:59.003 { 00:07:59.003 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.003 "dma_device_type": 2 00:07:59.003 } 00:07:59.003 ], 00:07:59.003 "driver_specific": {} 00:07:59.003 } 00:07:59.003 ] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.003 "name": "Existed_Raid", 00:07:59.003 "uuid": "a25aa725-e1ea-46fb-a9c3-b8af412913d6", 00:07:59.003 "strip_size_kb": 64, 00:07:59.003 "state": "online", 00:07:59.003 "raid_level": "concat", 00:07:59.003 "superblock": true, 00:07:59.003 "num_base_bdevs": 2, 00:07:59.003 "num_base_bdevs_discovered": 2, 00:07:59.003 "num_base_bdevs_operational": 2, 00:07:59.003 "base_bdevs_list": [ 00:07:59.003 { 00:07:59.003 "name": "BaseBdev1", 00:07:59.003 "uuid": "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29", 00:07:59.003 "is_configured": true, 00:07:59.003 "data_offset": 2048, 00:07:59.003 "data_size": 63488 00:07:59.003 }, 00:07:59.003 { 00:07:59.003 "name": "BaseBdev2", 00:07:59.003 "uuid": "15868bd1-02c6-4997-a975-6df08edded1d", 00:07:59.003 "is_configured": true, 00:07:59.003 "data_offset": 2048, 00:07:59.003 "data_size": 63488 00:07:59.003 } 00:07:59.003 ] 00:07:59.003 }' 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.003 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.574 [2024-10-29 10:57:04.830574] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:59.574 "name": "Existed_Raid", 00:07:59.574 "aliases": [ 00:07:59.574 "a25aa725-e1ea-46fb-a9c3-b8af412913d6" 00:07:59.574 ], 00:07:59.574 "product_name": "Raid Volume", 00:07:59.574 "block_size": 512, 00:07:59.574 "num_blocks": 126976, 00:07:59.574 "uuid": "a25aa725-e1ea-46fb-a9c3-b8af412913d6", 00:07:59.574 "assigned_rate_limits": { 00:07:59.574 "rw_ios_per_sec": 0, 00:07:59.574 "rw_mbytes_per_sec": 0, 00:07:59.574 "r_mbytes_per_sec": 0, 00:07:59.574 "w_mbytes_per_sec": 0 00:07:59.574 }, 00:07:59.574 "claimed": false, 00:07:59.574 "zoned": false, 00:07:59.574 "supported_io_types": { 00:07:59.574 "read": true, 00:07:59.574 "write": true, 00:07:59.574 "unmap": true, 00:07:59.574 "flush": true, 00:07:59.574 "reset": true, 00:07:59.574 "nvme_admin": false, 00:07:59.574 "nvme_io": false, 00:07:59.574 "nvme_io_md": false, 00:07:59.574 "write_zeroes": true, 00:07:59.574 "zcopy": false, 00:07:59.574 "get_zone_info": false, 00:07:59.574 "zone_management": false, 00:07:59.574 "zone_append": false, 00:07:59.574 "compare": false, 00:07:59.574 "compare_and_write": false, 00:07:59.574 "abort": false, 00:07:59.574 "seek_hole": false, 00:07:59.574 "seek_data": false, 00:07:59.574 "copy": false, 00:07:59.574 "nvme_iov_md": false 00:07:59.574 }, 00:07:59.574 "memory_domains": [ 00:07:59.574 { 00:07:59.574 "dma_device_id": "system", 00:07:59.574 "dma_device_type": 1 00:07:59.574 }, 00:07:59.574 { 00:07:59.574 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.574 "dma_device_type": 2 00:07:59.574 }, 00:07:59.574 { 00:07:59.574 "dma_device_id": "system", 00:07:59.574 "dma_device_type": 1 00:07:59.574 }, 00:07:59.574 { 00:07:59.574 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:59.574 "dma_device_type": 2 00:07:59.574 } 00:07:59.574 ], 00:07:59.574 "driver_specific": { 00:07:59.574 "raid": { 00:07:59.574 "uuid": "a25aa725-e1ea-46fb-a9c3-b8af412913d6", 00:07:59.574 "strip_size_kb": 64, 00:07:59.574 "state": "online", 00:07:59.574 "raid_level": "concat", 00:07:59.574 "superblock": true, 00:07:59.574 "num_base_bdevs": 2, 00:07:59.574 "num_base_bdevs_discovered": 2, 00:07:59.574 "num_base_bdevs_operational": 2, 00:07:59.574 "base_bdevs_list": [ 00:07:59.574 { 00:07:59.574 "name": "BaseBdev1", 00:07:59.574 "uuid": "8c8d9a15-0c68-4f29-8f8e-edbf661b3f29", 00:07:59.574 "is_configured": true, 00:07:59.574 "data_offset": 2048, 00:07:59.574 "data_size": 63488 00:07:59.574 }, 00:07:59.574 { 00:07:59.574 "name": "BaseBdev2", 00:07:59.574 "uuid": "15868bd1-02c6-4997-a975-6df08edded1d", 00:07:59.574 "is_configured": true, 00:07:59.574 "data_offset": 2048, 00:07:59.574 "data_size": 63488 00:07:59.574 } 00:07:59.574 ] 00:07:59.574 } 00:07:59.574 } 00:07:59.574 }' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:59.574 BaseBdev2' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.574 10:57:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.574 [2024-10-29 10:57:05.030005] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:59.574 [2024-10-29 10:57:05.030101] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:59.574 [2024-10-29 10:57:05.030193] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:59.574 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.575 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.835 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.835 "name": "Existed_Raid", 00:07:59.835 "uuid": "a25aa725-e1ea-46fb-a9c3-b8af412913d6", 00:07:59.835 "strip_size_kb": 64, 00:07:59.835 "state": "offline", 00:07:59.835 "raid_level": "concat", 00:07:59.835 "superblock": true, 00:07:59.835 "num_base_bdevs": 2, 00:07:59.835 "num_base_bdevs_discovered": 1, 00:07:59.835 "num_base_bdevs_operational": 1, 00:07:59.835 "base_bdevs_list": [ 00:07:59.835 { 00:07:59.835 "name": null, 00:07:59.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.835 "is_configured": false, 00:07:59.835 "data_offset": 0, 00:07:59.835 "data_size": 63488 00:07:59.835 }, 00:07:59.835 { 00:07:59.835 "name": "BaseBdev2", 00:07:59.835 "uuid": "15868bd1-02c6-4997-a975-6df08edded1d", 00:07:59.835 "is_configured": true, 00:07:59.835 "data_offset": 2048, 00:07:59.835 "data_size": 63488 00:07:59.835 } 00:07:59.835 ] 00:07:59.835 }' 00:07:59.835 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.835 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.096 [2024-10-29 10:57:05.464979] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:00.096 [2024-10-29 10:57:05.465076] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73436 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 73436 ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 73436 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73436 00:08:00.096 killing process with pid 73436 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73436' 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 73436 00:08:00.096 [2024-10-29 10:57:05.568417] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:00.096 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 73436 00:08:00.096 [2024-10-29 10:57:05.569428] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:00.356 10:57:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:00.356 00:08:00.356 real 0m3.722s 00:08:00.356 user 0m5.845s 00:08:00.356 sys 0m0.757s 00:08:00.356 ************************************ 00:08:00.356 END TEST raid_state_function_test_sb 00:08:00.356 ************************************ 00:08:00.356 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:00.356 10:57:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:00.356 10:57:05 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:08:00.356 10:57:05 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:08:00.356 10:57:05 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:00.356 10:57:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:00.356 ************************************ 00:08:00.356 START TEST raid_superblock_test 00:08:00.356 ************************************ 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test concat 2 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:00.616 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73666 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73666 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 73666 ']' 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:00.617 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:00.617 10:57:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.617 [2024-10-29 10:57:05.941535] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:00.617 [2024-10-29 10:57:05.941657] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73666 ] 00:08:00.617 [2024-10-29 10:57:06.091227] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:00.877 [2024-10-29 10:57:06.118003] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:00.877 [2024-10-29 10:57:06.162150] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:00.877 [2024-10-29 10:57:06.162179] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.455 malloc1 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.455 [2024-10-29 10:57:06.793635] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:01.455 [2024-10-29 10:57:06.793751] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:01.455 [2024-10-29 10:57:06.793788] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:01.455 [2024-10-29 10:57:06.793823] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:01.455 [2024-10-29 10:57:06.795943] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:01.455 [2024-10-29 10:57:06.796021] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:01.455 pt1 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.455 malloc2 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.455 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.456 [2024-10-29 10:57:06.826417] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:01.456 [2024-10-29 10:57:06.826531] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:01.456 [2024-10-29 10:57:06.826562] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:01.456 [2024-10-29 10:57:06.826591] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:01.456 [2024-10-29 10:57:06.828651] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:01.456 [2024-10-29 10:57:06.828741] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:01.456 pt2 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.456 [2024-10-29 10:57:06.838451] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:01.456 [2024-10-29 10:57:06.840282] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:01.456 [2024-10-29 10:57:06.840460] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:08:01.456 [2024-10-29 10:57:06.840476] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:01.456 [2024-10-29 10:57:06.840724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:01.456 [2024-10-29 10:57:06.840858] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:08:01.456 [2024-10-29 10:57:06.840868] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:08:01.456 [2024-10-29 10:57:06.841003] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.456 "name": "raid_bdev1", 00:08:01.456 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:01.456 "strip_size_kb": 64, 00:08:01.456 "state": "online", 00:08:01.456 "raid_level": "concat", 00:08:01.456 "superblock": true, 00:08:01.456 "num_base_bdevs": 2, 00:08:01.456 "num_base_bdevs_discovered": 2, 00:08:01.456 "num_base_bdevs_operational": 2, 00:08:01.456 "base_bdevs_list": [ 00:08:01.456 { 00:08:01.456 "name": "pt1", 00:08:01.456 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:01.456 "is_configured": true, 00:08:01.456 "data_offset": 2048, 00:08:01.456 "data_size": 63488 00:08:01.456 }, 00:08:01.456 { 00:08:01.456 "name": "pt2", 00:08:01.456 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.456 "is_configured": true, 00:08:01.456 "data_offset": 2048, 00:08:01.456 "data_size": 63488 00:08:01.456 } 00:08:01.456 ] 00:08:01.456 }' 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.456 10:57:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:02.026 [2024-10-29 10:57:07.277928] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:02.026 "name": "raid_bdev1", 00:08:02.026 "aliases": [ 00:08:02.026 "c86d4f63-85e9-4693-a16a-ef6eaf210e21" 00:08:02.026 ], 00:08:02.026 "product_name": "Raid Volume", 00:08:02.026 "block_size": 512, 00:08:02.026 "num_blocks": 126976, 00:08:02.026 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:02.026 "assigned_rate_limits": { 00:08:02.026 "rw_ios_per_sec": 0, 00:08:02.026 "rw_mbytes_per_sec": 0, 00:08:02.026 "r_mbytes_per_sec": 0, 00:08:02.026 "w_mbytes_per_sec": 0 00:08:02.026 }, 00:08:02.026 "claimed": false, 00:08:02.026 "zoned": false, 00:08:02.026 "supported_io_types": { 00:08:02.026 "read": true, 00:08:02.026 "write": true, 00:08:02.026 "unmap": true, 00:08:02.026 "flush": true, 00:08:02.026 "reset": true, 00:08:02.026 "nvme_admin": false, 00:08:02.026 "nvme_io": false, 00:08:02.026 "nvme_io_md": false, 00:08:02.026 "write_zeroes": true, 00:08:02.026 "zcopy": false, 00:08:02.026 "get_zone_info": false, 00:08:02.026 "zone_management": false, 00:08:02.026 "zone_append": false, 00:08:02.026 "compare": false, 00:08:02.026 "compare_and_write": false, 00:08:02.026 "abort": false, 00:08:02.026 "seek_hole": false, 00:08:02.026 "seek_data": false, 00:08:02.026 "copy": false, 00:08:02.026 "nvme_iov_md": false 00:08:02.026 }, 00:08:02.026 "memory_domains": [ 00:08:02.026 { 00:08:02.026 "dma_device_id": "system", 00:08:02.026 "dma_device_type": 1 00:08:02.026 }, 00:08:02.026 { 00:08:02.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:02.026 "dma_device_type": 2 00:08:02.026 }, 00:08:02.026 { 00:08:02.026 "dma_device_id": "system", 00:08:02.026 "dma_device_type": 1 00:08:02.026 }, 00:08:02.026 { 00:08:02.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:02.026 "dma_device_type": 2 00:08:02.026 } 00:08:02.026 ], 00:08:02.026 "driver_specific": { 00:08:02.026 "raid": { 00:08:02.026 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:02.026 "strip_size_kb": 64, 00:08:02.026 "state": "online", 00:08:02.026 "raid_level": "concat", 00:08:02.026 "superblock": true, 00:08:02.026 "num_base_bdevs": 2, 00:08:02.026 "num_base_bdevs_discovered": 2, 00:08:02.026 "num_base_bdevs_operational": 2, 00:08:02.026 "base_bdevs_list": [ 00:08:02.026 { 00:08:02.026 "name": "pt1", 00:08:02.026 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:02.026 "is_configured": true, 00:08:02.026 "data_offset": 2048, 00:08:02.026 "data_size": 63488 00:08:02.026 }, 00:08:02.026 { 00:08:02.026 "name": "pt2", 00:08:02.026 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.026 "is_configured": true, 00:08:02.026 "data_offset": 2048, 00:08:02.026 "data_size": 63488 00:08:02.026 } 00:08:02.026 ] 00:08:02.026 } 00:08:02.026 } 00:08:02.026 }' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:02.026 pt2' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.026 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.026 [2024-10-29 10:57:07.509443] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c86d4f63-85e9-4693-a16a-ef6eaf210e21 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z c86d4f63-85e9-4693-a16a-ef6eaf210e21 ']' 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.286 [2024-10-29 10:57:07.553132] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:02.286 [2024-10-29 10:57:07.553200] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:02.286 [2024-10-29 10:57:07.553291] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:02.286 [2024-10-29 10:57:07.553350] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:02.286 [2024-10-29 10:57:07.553377] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:02.286 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 [2024-10-29 10:57:07.684941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:02.287 [2024-10-29 10:57:07.686721] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:02.287 [2024-10-29 10:57:07.686785] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:02.287 [2024-10-29 10:57:07.686834] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:02.287 [2024-10-29 10:57:07.686849] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:02.287 [2024-10-29 10:57:07.686858] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:02.287 request: 00:08:02.287 { 00:08:02.287 "name": "raid_bdev1", 00:08:02.287 "raid_level": "concat", 00:08:02.287 "base_bdevs": [ 00:08:02.287 "malloc1", 00:08:02.287 "malloc2" 00:08:02.287 ], 00:08:02.287 "strip_size_kb": 64, 00:08:02.287 "superblock": false, 00:08:02.287 "method": "bdev_raid_create", 00:08:02.287 "req_id": 1 00:08:02.287 } 00:08:02.287 Got JSON-RPC error response 00:08:02.287 response: 00:08:02.287 { 00:08:02.287 "code": -17, 00:08:02.287 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:02.287 } 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 [2024-10-29 10:57:07.736801] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:02.287 [2024-10-29 10:57:07.736891] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.287 [2024-10-29 10:57:07.736913] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:02.287 [2024-10-29 10:57:07.736921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.287 [2024-10-29 10:57:07.739056] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.287 pt1 00:08:02.287 [2024-10-29 10:57:07.739136] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:02.287 [2024-10-29 10:57:07.739215] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:02.287 [2024-10-29 10:57:07.739279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.287 "name": "raid_bdev1", 00:08:02.287 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:02.287 "strip_size_kb": 64, 00:08:02.287 "state": "configuring", 00:08:02.287 "raid_level": "concat", 00:08:02.287 "superblock": true, 00:08:02.287 "num_base_bdevs": 2, 00:08:02.287 "num_base_bdevs_discovered": 1, 00:08:02.287 "num_base_bdevs_operational": 2, 00:08:02.287 "base_bdevs_list": [ 00:08:02.287 { 00:08:02.287 "name": "pt1", 00:08:02.287 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:02.287 "is_configured": true, 00:08:02.287 "data_offset": 2048, 00:08:02.287 "data_size": 63488 00:08:02.287 }, 00:08:02.287 { 00:08:02.287 "name": null, 00:08:02.287 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.287 "is_configured": false, 00:08:02.287 "data_offset": 2048, 00:08:02.287 "data_size": 63488 00:08:02.287 } 00:08:02.287 ] 00:08:02.287 }' 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.287 10:57:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.856 [2024-10-29 10:57:08.136162] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:02.856 [2024-10-29 10:57:08.136275] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.856 [2024-10-29 10:57:08.136317] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:02.856 [2024-10-29 10:57:08.136347] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.856 [2024-10-29 10:57:08.136848] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.856 [2024-10-29 10:57:08.136904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:02.856 [2024-10-29 10:57:08.137025] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:02.856 [2024-10-29 10:57:08.137086] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:02.856 [2024-10-29 10:57:08.137228] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:02.856 [2024-10-29 10:57:08.137266] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:02.856 [2024-10-29 10:57:08.137542] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:02.856 [2024-10-29 10:57:08.137704] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:02.856 [2024-10-29 10:57:08.137752] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:02.856 [2024-10-29 10:57:08.137932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:02.856 pt2 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.856 "name": "raid_bdev1", 00:08:02.856 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:02.856 "strip_size_kb": 64, 00:08:02.856 "state": "online", 00:08:02.856 "raid_level": "concat", 00:08:02.856 "superblock": true, 00:08:02.856 "num_base_bdevs": 2, 00:08:02.856 "num_base_bdevs_discovered": 2, 00:08:02.856 "num_base_bdevs_operational": 2, 00:08:02.856 "base_bdevs_list": [ 00:08:02.856 { 00:08:02.856 "name": "pt1", 00:08:02.856 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:02.856 "is_configured": true, 00:08:02.856 "data_offset": 2048, 00:08:02.856 "data_size": 63488 00:08:02.856 }, 00:08:02.856 { 00:08:02.856 "name": "pt2", 00:08:02.856 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.856 "is_configured": true, 00:08:02.856 "data_offset": 2048, 00:08:02.856 "data_size": 63488 00:08:02.856 } 00:08:02.856 ] 00:08:02.856 }' 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.856 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.115 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.115 [2024-10-29 10:57:08.611673] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:03.375 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.375 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:03.375 "name": "raid_bdev1", 00:08:03.375 "aliases": [ 00:08:03.375 "c86d4f63-85e9-4693-a16a-ef6eaf210e21" 00:08:03.375 ], 00:08:03.375 "product_name": "Raid Volume", 00:08:03.375 "block_size": 512, 00:08:03.375 "num_blocks": 126976, 00:08:03.375 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:03.375 "assigned_rate_limits": { 00:08:03.375 "rw_ios_per_sec": 0, 00:08:03.375 "rw_mbytes_per_sec": 0, 00:08:03.375 "r_mbytes_per_sec": 0, 00:08:03.375 "w_mbytes_per_sec": 0 00:08:03.375 }, 00:08:03.375 "claimed": false, 00:08:03.375 "zoned": false, 00:08:03.375 "supported_io_types": { 00:08:03.375 "read": true, 00:08:03.375 "write": true, 00:08:03.375 "unmap": true, 00:08:03.375 "flush": true, 00:08:03.375 "reset": true, 00:08:03.375 "nvme_admin": false, 00:08:03.375 "nvme_io": false, 00:08:03.375 "nvme_io_md": false, 00:08:03.375 "write_zeroes": true, 00:08:03.375 "zcopy": false, 00:08:03.375 "get_zone_info": false, 00:08:03.375 "zone_management": false, 00:08:03.375 "zone_append": false, 00:08:03.375 "compare": false, 00:08:03.375 "compare_and_write": false, 00:08:03.375 "abort": false, 00:08:03.375 "seek_hole": false, 00:08:03.375 "seek_data": false, 00:08:03.375 "copy": false, 00:08:03.375 "nvme_iov_md": false 00:08:03.375 }, 00:08:03.375 "memory_domains": [ 00:08:03.375 { 00:08:03.375 "dma_device_id": "system", 00:08:03.375 "dma_device_type": 1 00:08:03.375 }, 00:08:03.375 { 00:08:03.375 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.376 "dma_device_type": 2 00:08:03.376 }, 00:08:03.376 { 00:08:03.376 "dma_device_id": "system", 00:08:03.376 "dma_device_type": 1 00:08:03.376 }, 00:08:03.376 { 00:08:03.376 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.376 "dma_device_type": 2 00:08:03.376 } 00:08:03.376 ], 00:08:03.376 "driver_specific": { 00:08:03.376 "raid": { 00:08:03.376 "uuid": "c86d4f63-85e9-4693-a16a-ef6eaf210e21", 00:08:03.376 "strip_size_kb": 64, 00:08:03.376 "state": "online", 00:08:03.376 "raid_level": "concat", 00:08:03.376 "superblock": true, 00:08:03.376 "num_base_bdevs": 2, 00:08:03.376 "num_base_bdevs_discovered": 2, 00:08:03.376 "num_base_bdevs_operational": 2, 00:08:03.376 "base_bdevs_list": [ 00:08:03.376 { 00:08:03.376 "name": "pt1", 00:08:03.376 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:03.376 "is_configured": true, 00:08:03.376 "data_offset": 2048, 00:08:03.376 "data_size": 63488 00:08:03.376 }, 00:08:03.376 { 00:08:03.376 "name": "pt2", 00:08:03.376 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:03.376 "is_configured": true, 00:08:03.376 "data_offset": 2048, 00:08:03.376 "data_size": 63488 00:08:03.376 } 00:08:03.376 ] 00:08:03.376 } 00:08:03.376 } 00:08:03.376 }' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:03.376 pt2' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.376 [2024-10-29 10:57:08.827459] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' c86d4f63-85e9-4693-a16a-ef6eaf210e21 '!=' c86d4f63-85e9-4693-a16a-ef6eaf210e21 ']' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73666 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 73666 ']' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 73666 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:03.376 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73666 00:08:03.635 killing process with pid 73666 00:08:03.635 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:03.635 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:03.635 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73666' 00:08:03.635 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 73666 00:08:03.635 [2024-10-29 10:57:08.880199] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:03.635 [2024-10-29 10:57:08.880277] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:03.635 [2024-10-29 10:57:08.880331] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:03.635 [2024-10-29 10:57:08.880340] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:03.635 10:57:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 73666 00:08:03.635 [2024-10-29 10:57:08.902660] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:03.635 10:57:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:03.635 00:08:03.635 real 0m3.261s 00:08:03.635 user 0m5.059s 00:08:03.635 sys 0m0.668s 00:08:03.635 10:57:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:03.635 ************************************ 00:08:03.635 END TEST raid_superblock_test 00:08:03.635 ************************************ 00:08:03.635 10:57:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.894 10:57:09 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:08:03.894 10:57:09 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:03.894 10:57:09 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:03.894 10:57:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:03.894 ************************************ 00:08:03.894 START TEST raid_read_error_test 00:08:03.894 ************************************ 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 2 read 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4IyQ6hN4sK 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73867 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73867 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 73867 ']' 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:03.894 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:03.894 10:57:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.894 [2024-10-29 10:57:09.282780] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:03.894 [2024-10-29 10:57:09.282902] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73867 ] 00:08:04.153 [2024-10-29 10:57:09.453779] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:04.153 [2024-10-29 10:57:09.479329] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:04.153 [2024-10-29 10:57:09.522490] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:04.153 [2024-10-29 10:57:09.522522] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 BaseBdev1_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 true 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 [2024-10-29 10:57:10.149117] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:04.721 [2024-10-29 10:57:10.149170] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:04.721 [2024-10-29 10:57:10.149188] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:04.721 [2024-10-29 10:57:10.149197] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:04.721 [2024-10-29 10:57:10.151280] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:04.721 [2024-10-29 10:57:10.151318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:04.721 BaseBdev1 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 BaseBdev2_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 true 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 [2024-10-29 10:57:10.189686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:04.721 [2024-10-29 10:57:10.189734] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:04.721 [2024-10-29 10:57:10.189753] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:04.721 [2024-10-29 10:57:10.189761] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:04.721 [2024-10-29 10:57:10.191782] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:04.721 [2024-10-29 10:57:10.191877] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:04.721 BaseBdev2 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.721 [2024-10-29 10:57:10.201723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:04.721 [2024-10-29 10:57:10.203565] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:04.721 [2024-10-29 10:57:10.203726] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:04.721 [2024-10-29 10:57:10.203739] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:04.721 [2024-10-29 10:57:10.203987] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:04.721 [2024-10-29 10:57:10.204109] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:04.721 [2024-10-29 10:57:10.204121] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:04.721 [2024-10-29 10:57:10.204242] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.721 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.980 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.980 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.980 "name": "raid_bdev1", 00:08:04.980 "uuid": "7e85ba33-3525-4cc8-94ad-4a6d437cd935", 00:08:04.980 "strip_size_kb": 64, 00:08:04.980 "state": "online", 00:08:04.980 "raid_level": "concat", 00:08:04.980 "superblock": true, 00:08:04.980 "num_base_bdevs": 2, 00:08:04.980 "num_base_bdevs_discovered": 2, 00:08:04.980 "num_base_bdevs_operational": 2, 00:08:04.980 "base_bdevs_list": [ 00:08:04.980 { 00:08:04.980 "name": "BaseBdev1", 00:08:04.980 "uuid": "68f4fd29-2353-5156-8ed5-d1d38253761c", 00:08:04.980 "is_configured": true, 00:08:04.980 "data_offset": 2048, 00:08:04.980 "data_size": 63488 00:08:04.980 }, 00:08:04.980 { 00:08:04.980 "name": "BaseBdev2", 00:08:04.980 "uuid": "d5cc27f4-40db-5afb-aed1-79689260edc2", 00:08:04.980 "is_configured": true, 00:08:04.980 "data_offset": 2048, 00:08:04.980 "data_size": 63488 00:08:04.980 } 00:08:04.980 ] 00:08:04.980 }' 00:08:04.980 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.980 10:57:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.239 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:05.239 10:57:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:05.497 [2024-10-29 10:57:10.745202] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.436 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.436 "name": "raid_bdev1", 00:08:06.436 "uuid": "7e85ba33-3525-4cc8-94ad-4a6d437cd935", 00:08:06.436 "strip_size_kb": 64, 00:08:06.436 "state": "online", 00:08:06.436 "raid_level": "concat", 00:08:06.436 "superblock": true, 00:08:06.436 "num_base_bdevs": 2, 00:08:06.436 "num_base_bdevs_discovered": 2, 00:08:06.436 "num_base_bdevs_operational": 2, 00:08:06.436 "base_bdevs_list": [ 00:08:06.436 { 00:08:06.436 "name": "BaseBdev1", 00:08:06.436 "uuid": "68f4fd29-2353-5156-8ed5-d1d38253761c", 00:08:06.436 "is_configured": true, 00:08:06.436 "data_offset": 2048, 00:08:06.436 "data_size": 63488 00:08:06.437 }, 00:08:06.437 { 00:08:06.437 "name": "BaseBdev2", 00:08:06.437 "uuid": "d5cc27f4-40db-5afb-aed1-79689260edc2", 00:08:06.437 "is_configured": true, 00:08:06.437 "data_offset": 2048, 00:08:06.437 "data_size": 63488 00:08:06.437 } 00:08:06.437 ] 00:08:06.437 }' 00:08:06.437 10:57:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.437 10:57:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.698 [2024-10-29 10:57:12.113010] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:06.698 [2024-10-29 10:57:12.113046] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:06.698 [2024-10-29 10:57:12.115604] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:06.698 [2024-10-29 10:57:12.115723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:06.698 [2024-10-29 10:57:12.115766] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:06.698 [2024-10-29 10:57:12.115776] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:06.698 { 00:08:06.698 "results": [ 00:08:06.698 { 00:08:06.698 "job": "raid_bdev1", 00:08:06.698 "core_mask": "0x1", 00:08:06.698 "workload": "randrw", 00:08:06.698 "percentage": 50, 00:08:06.698 "status": "finished", 00:08:06.698 "queue_depth": 1, 00:08:06.698 "io_size": 131072, 00:08:06.698 "runtime": 1.368553, 00:08:06.698 "iops": 17282.487415540356, 00:08:06.698 "mibps": 2160.3109269425445, 00:08:06.698 "io_failed": 1, 00:08:06.698 "io_timeout": 0, 00:08:06.698 "avg_latency_us": 79.98324951901925, 00:08:06.698 "min_latency_us": 25.4882096069869, 00:08:06.698 "max_latency_us": 1366.5257641921398 00:08:06.698 } 00:08:06.698 ], 00:08:06.698 "core_count": 1 00:08:06.698 } 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73867 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 73867 ']' 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 73867 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73867 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73867' 00:08:06.698 killing process with pid 73867 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 73867 00:08:06.698 [2024-10-29 10:57:12.167353] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:06.698 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 73867 00:08:06.698 [2024-10-29 10:57:12.182849] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4IyQ6hN4sK 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:06.958 00:08:06.958 real 0m3.215s 00:08:06.958 user 0m4.105s 00:08:06.958 sys 0m0.518s 00:08:06.958 ************************************ 00:08:06.958 END TEST raid_read_error_test 00:08:06.958 ************************************ 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:06.958 10:57:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.958 10:57:12 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:08:06.958 10:57:12 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:06.958 10:57:12 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:06.958 10:57:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:07.217 ************************************ 00:08:07.217 START TEST raid_write_error_test 00:08:07.217 ************************************ 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 2 write 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.PuXuxpxpT8 00:08:07.217 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73996 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73996 00:08:07.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 73996 ']' 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:07.218 10:57:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.218 [2024-10-29 10:57:12.570394] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:07.218 [2024-10-29 10:57:12.570517] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73996 ] 00:08:07.477 [2024-10-29 10:57:12.742348] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:07.477 [2024-10-29 10:57:12.767905] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:07.477 [2024-10-29 10:57:12.811317] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:07.477 [2024-10-29 10:57:12.811356] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 BaseBdev1_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 true 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 [2024-10-29 10:57:13.429843] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:08.049 [2024-10-29 10:57:13.429893] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.049 [2024-10-29 10:57:13.429912] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:08.049 [2024-10-29 10:57:13.429921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.049 [2024-10-29 10:57:13.432016] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.049 [2024-10-29 10:57:13.432053] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:08.049 BaseBdev1 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 BaseBdev2_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 true 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 [2024-10-29 10:57:13.470602] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:08.049 [2024-10-29 10:57:13.470693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.049 [2024-10-29 10:57:13.470716] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:08.049 [2024-10-29 10:57:13.470724] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.049 [2024-10-29 10:57:13.472821] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.049 [2024-10-29 10:57:13.472857] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:08.049 BaseBdev2 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.049 [2024-10-29 10:57:13.482631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:08.049 [2024-10-29 10:57:13.484498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:08.049 [2024-10-29 10:57:13.484665] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:08.049 [2024-10-29 10:57:13.484683] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:08.049 [2024-10-29 10:57:13.484927] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:08.049 [2024-10-29 10:57:13.485069] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:08.049 [2024-10-29 10:57:13.485082] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:08.049 [2024-10-29 10:57:13.485204] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.049 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.050 "name": "raid_bdev1", 00:08:08.050 "uuid": "225c8c74-0de6-4a3a-94d1-5b11731e8bb3", 00:08:08.050 "strip_size_kb": 64, 00:08:08.050 "state": "online", 00:08:08.050 "raid_level": "concat", 00:08:08.050 "superblock": true, 00:08:08.050 "num_base_bdevs": 2, 00:08:08.050 "num_base_bdevs_discovered": 2, 00:08:08.050 "num_base_bdevs_operational": 2, 00:08:08.050 "base_bdevs_list": [ 00:08:08.050 { 00:08:08.050 "name": "BaseBdev1", 00:08:08.050 "uuid": "0a85f71d-b99f-5507-b515-6c132ab3bbe6", 00:08:08.050 "is_configured": true, 00:08:08.050 "data_offset": 2048, 00:08:08.050 "data_size": 63488 00:08:08.050 }, 00:08:08.050 { 00:08:08.050 "name": "BaseBdev2", 00:08:08.050 "uuid": "dc9adb4f-dbb4-5c03-a6e7-35e7bb7f91e7", 00:08:08.050 "is_configured": true, 00:08:08.050 "data_offset": 2048, 00:08:08.050 "data_size": 63488 00:08:08.050 } 00:08:08.050 ] 00:08:08.050 }' 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.050 10:57:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.620 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:08.620 10:57:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:08.620 [2024-10-29 10:57:13.998384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:09.561 "name": "raid_bdev1", 00:08:09.561 "uuid": "225c8c74-0de6-4a3a-94d1-5b11731e8bb3", 00:08:09.561 "strip_size_kb": 64, 00:08:09.561 "state": "online", 00:08:09.561 "raid_level": "concat", 00:08:09.561 "superblock": true, 00:08:09.561 "num_base_bdevs": 2, 00:08:09.561 "num_base_bdevs_discovered": 2, 00:08:09.561 "num_base_bdevs_operational": 2, 00:08:09.561 "base_bdevs_list": [ 00:08:09.561 { 00:08:09.561 "name": "BaseBdev1", 00:08:09.561 "uuid": "0a85f71d-b99f-5507-b515-6c132ab3bbe6", 00:08:09.561 "is_configured": true, 00:08:09.561 "data_offset": 2048, 00:08:09.561 "data_size": 63488 00:08:09.561 }, 00:08:09.561 { 00:08:09.561 "name": "BaseBdev2", 00:08:09.561 "uuid": "dc9adb4f-dbb4-5c03-a6e7-35e7bb7f91e7", 00:08:09.561 "is_configured": true, 00:08:09.561 "data_offset": 2048, 00:08:09.561 "data_size": 63488 00:08:09.561 } 00:08:09.561 ] 00:08:09.561 }' 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:09.561 10:57:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.132 [2024-10-29 10:57:15.346734] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:10.132 [2024-10-29 10:57:15.346769] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:10.132 [2024-10-29 10:57:15.349290] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:10.132 [2024-10-29 10:57:15.349334] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:10.132 [2024-10-29 10:57:15.349384] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:10.132 [2024-10-29 10:57:15.349395] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:10.132 { 00:08:10.132 "results": [ 00:08:10.132 { 00:08:10.132 "job": "raid_bdev1", 00:08:10.132 "core_mask": "0x1", 00:08:10.132 "workload": "randrw", 00:08:10.132 "percentage": 50, 00:08:10.132 "status": "finished", 00:08:10.132 "queue_depth": 1, 00:08:10.132 "io_size": 131072, 00:08:10.132 "runtime": 1.348516, 00:08:10.132 "iops": 15458.47435254754, 00:08:10.132 "mibps": 1932.3092940684426, 00:08:10.132 "io_failed": 1, 00:08:10.132 "io_timeout": 0, 00:08:10.132 "avg_latency_us": 90.1327109573325, 00:08:10.132 "min_latency_us": 24.482096069868994, 00:08:10.132 "max_latency_us": 1459.5353711790392 00:08:10.132 } 00:08:10.132 ], 00:08:10.132 "core_count": 1 00:08:10.132 } 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73996 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 73996 ']' 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 73996 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 73996 00:08:10.132 killing process with pid 73996 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 73996' 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 73996 00:08:10.132 [2024-10-29 10:57:15.395801] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 73996 00:08:10.132 [2024-10-29 10:57:15.411941] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.PuXuxpxpT8 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:10.132 00:08:10.132 real 0m3.157s 00:08:10.132 user 0m4.024s 00:08:10.132 sys 0m0.496s 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:10.132 ************************************ 00:08:10.132 END TEST raid_write_error_test 00:08:10.132 ************************************ 00:08:10.132 10:57:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.393 10:57:15 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:10.393 10:57:15 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:08:10.393 10:57:15 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:10.393 10:57:15 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:10.393 10:57:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:10.393 ************************************ 00:08:10.393 START TEST raid_state_function_test 00:08:10.393 ************************************ 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 2 false 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74123 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74123' 00:08:10.393 Process raid pid: 74123 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74123 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 74123 ']' 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:10.393 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:10.393 10:57:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.393 [2024-10-29 10:57:15.787404] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:10.393 [2024-10-29 10:57:15.787597] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:10.653 [2024-10-29 10:57:15.959081] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.653 [2024-10-29 10:57:15.985482] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:10.653 [2024-10-29 10:57:16.028248] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.653 [2024-10-29 10:57:16.028382] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.224 [2024-10-29 10:57:16.629248] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:11.224 [2024-10-29 10:57:16.629375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:11.224 [2024-10-29 10:57:16.629425] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.224 [2024-10-29 10:57:16.629450] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.224 "name": "Existed_Raid", 00:08:11.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.224 "strip_size_kb": 0, 00:08:11.224 "state": "configuring", 00:08:11.224 "raid_level": "raid1", 00:08:11.224 "superblock": false, 00:08:11.224 "num_base_bdevs": 2, 00:08:11.224 "num_base_bdevs_discovered": 0, 00:08:11.224 "num_base_bdevs_operational": 2, 00:08:11.224 "base_bdevs_list": [ 00:08:11.224 { 00:08:11.224 "name": "BaseBdev1", 00:08:11.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.224 "is_configured": false, 00:08:11.224 "data_offset": 0, 00:08:11.224 "data_size": 0 00:08:11.224 }, 00:08:11.224 { 00:08:11.224 "name": "BaseBdev2", 00:08:11.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.224 "is_configured": false, 00:08:11.224 "data_offset": 0, 00:08:11.224 "data_size": 0 00:08:11.224 } 00:08:11.224 ] 00:08:11.224 }' 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.224 10:57:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.815 [2024-10-29 10:57:17.040494] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:11.815 [2024-10-29 10:57:17.040591] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.815 [2024-10-29 10:57:17.052473] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:11.815 [2024-10-29 10:57:17.052513] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:11.815 [2024-10-29 10:57:17.052521] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.815 [2024-10-29 10:57:17.052545] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.815 [2024-10-29 10:57:17.073016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:11.815 BaseBdev1 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:11.815 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.816 [ 00:08:11.816 { 00:08:11.816 "name": "BaseBdev1", 00:08:11.816 "aliases": [ 00:08:11.816 "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49" 00:08:11.816 ], 00:08:11.816 "product_name": "Malloc disk", 00:08:11.816 "block_size": 512, 00:08:11.816 "num_blocks": 65536, 00:08:11.816 "uuid": "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49", 00:08:11.816 "assigned_rate_limits": { 00:08:11.816 "rw_ios_per_sec": 0, 00:08:11.816 "rw_mbytes_per_sec": 0, 00:08:11.816 "r_mbytes_per_sec": 0, 00:08:11.816 "w_mbytes_per_sec": 0 00:08:11.816 }, 00:08:11.816 "claimed": true, 00:08:11.816 "claim_type": "exclusive_write", 00:08:11.816 "zoned": false, 00:08:11.816 "supported_io_types": { 00:08:11.816 "read": true, 00:08:11.816 "write": true, 00:08:11.816 "unmap": true, 00:08:11.816 "flush": true, 00:08:11.816 "reset": true, 00:08:11.816 "nvme_admin": false, 00:08:11.816 "nvme_io": false, 00:08:11.816 "nvme_io_md": false, 00:08:11.816 "write_zeroes": true, 00:08:11.816 "zcopy": true, 00:08:11.816 "get_zone_info": false, 00:08:11.816 "zone_management": false, 00:08:11.816 "zone_append": false, 00:08:11.816 "compare": false, 00:08:11.816 "compare_and_write": false, 00:08:11.816 "abort": true, 00:08:11.816 "seek_hole": false, 00:08:11.816 "seek_data": false, 00:08:11.816 "copy": true, 00:08:11.816 "nvme_iov_md": false 00:08:11.816 }, 00:08:11.816 "memory_domains": [ 00:08:11.816 { 00:08:11.816 "dma_device_id": "system", 00:08:11.816 "dma_device_type": 1 00:08:11.816 }, 00:08:11.816 { 00:08:11.816 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:11.816 "dma_device_type": 2 00:08:11.816 } 00:08:11.816 ], 00:08:11.816 "driver_specific": {} 00:08:11.816 } 00:08:11.816 ] 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.816 "name": "Existed_Raid", 00:08:11.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.816 "strip_size_kb": 0, 00:08:11.816 "state": "configuring", 00:08:11.816 "raid_level": "raid1", 00:08:11.816 "superblock": false, 00:08:11.816 "num_base_bdevs": 2, 00:08:11.816 "num_base_bdevs_discovered": 1, 00:08:11.816 "num_base_bdevs_operational": 2, 00:08:11.816 "base_bdevs_list": [ 00:08:11.816 { 00:08:11.816 "name": "BaseBdev1", 00:08:11.816 "uuid": "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49", 00:08:11.816 "is_configured": true, 00:08:11.816 "data_offset": 0, 00:08:11.816 "data_size": 65536 00:08:11.816 }, 00:08:11.816 { 00:08:11.816 "name": "BaseBdev2", 00:08:11.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.816 "is_configured": false, 00:08:11.816 "data_offset": 0, 00:08:11.816 "data_size": 0 00:08:11.816 } 00:08:11.816 ] 00:08:11.816 }' 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.816 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.097 [2024-10-29 10:57:17.536242] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:12.097 [2024-10-29 10:57:17.536294] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.097 [2024-10-29 10:57:17.548234] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:12.097 [2024-10-29 10:57:17.550160] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:12.097 [2024-10-29 10:57:17.550250] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.097 "name": "Existed_Raid", 00:08:12.097 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.097 "strip_size_kb": 0, 00:08:12.097 "state": "configuring", 00:08:12.097 "raid_level": "raid1", 00:08:12.097 "superblock": false, 00:08:12.097 "num_base_bdevs": 2, 00:08:12.097 "num_base_bdevs_discovered": 1, 00:08:12.097 "num_base_bdevs_operational": 2, 00:08:12.097 "base_bdevs_list": [ 00:08:12.097 { 00:08:12.097 "name": "BaseBdev1", 00:08:12.097 "uuid": "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49", 00:08:12.097 "is_configured": true, 00:08:12.097 "data_offset": 0, 00:08:12.097 "data_size": 65536 00:08:12.097 }, 00:08:12.097 { 00:08:12.097 "name": "BaseBdev2", 00:08:12.097 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.097 "is_configured": false, 00:08:12.097 "data_offset": 0, 00:08:12.097 "data_size": 0 00:08:12.097 } 00:08:12.097 ] 00:08:12.097 }' 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.097 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.668 [2024-10-29 10:57:17.986354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:12.668 [2024-10-29 10:57:17.986417] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:12.668 [2024-10-29 10:57:17.986426] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:12.668 [2024-10-29 10:57:17.986712] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:12.668 [2024-10-29 10:57:17.986860] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:12.668 [2024-10-29 10:57:17.986875] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:12.668 [2024-10-29 10:57:17.987089] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:12.668 BaseBdev2 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.668 10:57:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.668 [ 00:08:12.668 { 00:08:12.668 "name": "BaseBdev2", 00:08:12.668 "aliases": [ 00:08:12.668 "7ff51e68-418b-4adf-8d02-aee25389e3ed" 00:08:12.668 ], 00:08:12.668 "product_name": "Malloc disk", 00:08:12.668 "block_size": 512, 00:08:12.668 "num_blocks": 65536, 00:08:12.668 "uuid": "7ff51e68-418b-4adf-8d02-aee25389e3ed", 00:08:12.668 "assigned_rate_limits": { 00:08:12.668 "rw_ios_per_sec": 0, 00:08:12.668 "rw_mbytes_per_sec": 0, 00:08:12.668 "r_mbytes_per_sec": 0, 00:08:12.668 "w_mbytes_per_sec": 0 00:08:12.668 }, 00:08:12.668 "claimed": true, 00:08:12.668 "claim_type": "exclusive_write", 00:08:12.668 "zoned": false, 00:08:12.668 "supported_io_types": { 00:08:12.668 "read": true, 00:08:12.668 "write": true, 00:08:12.668 "unmap": true, 00:08:12.668 "flush": true, 00:08:12.668 "reset": true, 00:08:12.668 "nvme_admin": false, 00:08:12.668 "nvme_io": false, 00:08:12.668 "nvme_io_md": false, 00:08:12.668 "write_zeroes": true, 00:08:12.668 "zcopy": true, 00:08:12.668 "get_zone_info": false, 00:08:12.668 "zone_management": false, 00:08:12.668 "zone_append": false, 00:08:12.668 "compare": false, 00:08:12.668 "compare_and_write": false, 00:08:12.668 "abort": true, 00:08:12.668 "seek_hole": false, 00:08:12.668 "seek_data": false, 00:08:12.668 "copy": true, 00:08:12.668 "nvme_iov_md": false 00:08:12.668 }, 00:08:12.668 "memory_domains": [ 00:08:12.668 { 00:08:12.668 "dma_device_id": "system", 00:08:12.668 "dma_device_type": 1 00:08:12.668 }, 00:08:12.668 { 00:08:12.668 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.668 "dma_device_type": 2 00:08:12.668 } 00:08:12.668 ], 00:08:12.668 "driver_specific": {} 00:08:12.668 } 00:08:12.668 ] 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.668 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.669 "name": "Existed_Raid", 00:08:12.669 "uuid": "304126ae-2a2e-4bbc-b8df-bcdfa1282214", 00:08:12.669 "strip_size_kb": 0, 00:08:12.669 "state": "online", 00:08:12.669 "raid_level": "raid1", 00:08:12.669 "superblock": false, 00:08:12.669 "num_base_bdevs": 2, 00:08:12.669 "num_base_bdevs_discovered": 2, 00:08:12.669 "num_base_bdevs_operational": 2, 00:08:12.669 "base_bdevs_list": [ 00:08:12.669 { 00:08:12.669 "name": "BaseBdev1", 00:08:12.669 "uuid": "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49", 00:08:12.669 "is_configured": true, 00:08:12.669 "data_offset": 0, 00:08:12.669 "data_size": 65536 00:08:12.669 }, 00:08:12.669 { 00:08:12.669 "name": "BaseBdev2", 00:08:12.669 "uuid": "7ff51e68-418b-4adf-8d02-aee25389e3ed", 00:08:12.669 "is_configured": true, 00:08:12.669 "data_offset": 0, 00:08:12.669 "data_size": 65536 00:08:12.669 } 00:08:12.669 ] 00:08:12.669 }' 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.669 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.240 [2024-10-29 10:57:18.481817] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:13.240 "name": "Existed_Raid", 00:08:13.240 "aliases": [ 00:08:13.240 "304126ae-2a2e-4bbc-b8df-bcdfa1282214" 00:08:13.240 ], 00:08:13.240 "product_name": "Raid Volume", 00:08:13.240 "block_size": 512, 00:08:13.240 "num_blocks": 65536, 00:08:13.240 "uuid": "304126ae-2a2e-4bbc-b8df-bcdfa1282214", 00:08:13.240 "assigned_rate_limits": { 00:08:13.240 "rw_ios_per_sec": 0, 00:08:13.240 "rw_mbytes_per_sec": 0, 00:08:13.240 "r_mbytes_per_sec": 0, 00:08:13.240 "w_mbytes_per_sec": 0 00:08:13.240 }, 00:08:13.240 "claimed": false, 00:08:13.240 "zoned": false, 00:08:13.240 "supported_io_types": { 00:08:13.240 "read": true, 00:08:13.240 "write": true, 00:08:13.240 "unmap": false, 00:08:13.240 "flush": false, 00:08:13.240 "reset": true, 00:08:13.240 "nvme_admin": false, 00:08:13.240 "nvme_io": false, 00:08:13.240 "nvme_io_md": false, 00:08:13.240 "write_zeroes": true, 00:08:13.240 "zcopy": false, 00:08:13.240 "get_zone_info": false, 00:08:13.240 "zone_management": false, 00:08:13.240 "zone_append": false, 00:08:13.240 "compare": false, 00:08:13.240 "compare_and_write": false, 00:08:13.240 "abort": false, 00:08:13.240 "seek_hole": false, 00:08:13.240 "seek_data": false, 00:08:13.240 "copy": false, 00:08:13.240 "nvme_iov_md": false 00:08:13.240 }, 00:08:13.240 "memory_domains": [ 00:08:13.240 { 00:08:13.240 "dma_device_id": "system", 00:08:13.240 "dma_device_type": 1 00:08:13.240 }, 00:08:13.240 { 00:08:13.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.240 "dma_device_type": 2 00:08:13.240 }, 00:08:13.240 { 00:08:13.240 "dma_device_id": "system", 00:08:13.240 "dma_device_type": 1 00:08:13.240 }, 00:08:13.240 { 00:08:13.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.240 "dma_device_type": 2 00:08:13.240 } 00:08:13.240 ], 00:08:13.240 "driver_specific": { 00:08:13.240 "raid": { 00:08:13.240 "uuid": "304126ae-2a2e-4bbc-b8df-bcdfa1282214", 00:08:13.240 "strip_size_kb": 0, 00:08:13.240 "state": "online", 00:08:13.240 "raid_level": "raid1", 00:08:13.240 "superblock": false, 00:08:13.240 "num_base_bdevs": 2, 00:08:13.240 "num_base_bdevs_discovered": 2, 00:08:13.240 "num_base_bdevs_operational": 2, 00:08:13.240 "base_bdevs_list": [ 00:08:13.240 { 00:08:13.240 "name": "BaseBdev1", 00:08:13.240 "uuid": "94ec2fdd-eaa5-40c2-a372-2f709a3a2b49", 00:08:13.240 "is_configured": true, 00:08:13.240 "data_offset": 0, 00:08:13.240 "data_size": 65536 00:08:13.240 }, 00:08:13.240 { 00:08:13.240 "name": "BaseBdev2", 00:08:13.240 "uuid": "7ff51e68-418b-4adf-8d02-aee25389e3ed", 00:08:13.240 "is_configured": true, 00:08:13.240 "data_offset": 0, 00:08:13.240 "data_size": 65536 00:08:13.240 } 00:08:13.240 ] 00:08:13.240 } 00:08:13.240 } 00:08:13.240 }' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:13.240 BaseBdev2' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.240 [2024-10-29 10:57:18.705204] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:13.240 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.241 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.501 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.501 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.501 "name": "Existed_Raid", 00:08:13.501 "uuid": "304126ae-2a2e-4bbc-b8df-bcdfa1282214", 00:08:13.501 "strip_size_kb": 0, 00:08:13.501 "state": "online", 00:08:13.501 "raid_level": "raid1", 00:08:13.501 "superblock": false, 00:08:13.501 "num_base_bdevs": 2, 00:08:13.501 "num_base_bdevs_discovered": 1, 00:08:13.501 "num_base_bdevs_operational": 1, 00:08:13.501 "base_bdevs_list": [ 00:08:13.501 { 00:08:13.501 "name": null, 00:08:13.501 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:13.501 "is_configured": false, 00:08:13.501 "data_offset": 0, 00:08:13.501 "data_size": 65536 00:08:13.501 }, 00:08:13.501 { 00:08:13.501 "name": "BaseBdev2", 00:08:13.501 "uuid": "7ff51e68-418b-4adf-8d02-aee25389e3ed", 00:08:13.501 "is_configured": true, 00:08:13.501 "data_offset": 0, 00:08:13.501 "data_size": 65536 00:08:13.501 } 00:08:13.501 ] 00:08:13.501 }' 00:08:13.501 10:57:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.501 10:57:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.762 [2024-10-29 10:57:19.195655] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:13.762 [2024-10-29 10:57:19.195762] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:13.762 [2024-10-29 10:57:19.206956] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:13.762 [2024-10-29 10:57:19.207003] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:13.762 [2024-10-29 10:57:19.207016] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:13.762 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74123 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 74123 ']' 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 74123 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 74123 00:08:14.023 killing process with pid 74123 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 74123' 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 74123 00:08:14.023 [2024-10-29 10:57:19.296712] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 74123 00:08:14.023 [2024-10-29 10:57:19.297706] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:14.023 ************************************ 00:08:14.023 END TEST raid_state_function_test 00:08:14.023 ************************************ 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:14.023 00:08:14.023 real 0m3.815s 00:08:14.023 user 0m6.004s 00:08:14.023 sys 0m0.789s 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:14.023 10:57:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.284 10:57:19 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:08:14.284 10:57:19 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:14.284 10:57:19 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:14.284 10:57:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:14.284 ************************************ 00:08:14.284 START TEST raid_state_function_test_sb 00:08:14.284 ************************************ 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 2 true 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74365 00:08:14.284 Process raid pid: 74365 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74365' 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74365 00:08:14.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 74365 ']' 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:14.284 10:57:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:14.284 [2024-10-29 10:57:19.673187] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:14.284 [2024-10-29 10:57:19.673331] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:14.543 [2024-10-29 10:57:19.845435] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:14.543 [2024-10-29 10:57:19.870681] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.543 [2024-10-29 10:57:19.912076] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:14.543 [2024-10-29 10:57:19.912110] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.111 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:15.111 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:08:15.111 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:15.111 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.111 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.111 [2024-10-29 10:57:20.521084] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:15.111 [2024-10-29 10:57:20.521137] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:15.112 [2024-10-29 10:57:20.521151] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:15.112 [2024-10-29 10:57:20.521161] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.112 "name": "Existed_Raid", 00:08:15.112 "uuid": "ebce8f27-0578-4e62-968e-ac89d3325cb2", 00:08:15.112 "strip_size_kb": 0, 00:08:15.112 "state": "configuring", 00:08:15.112 "raid_level": "raid1", 00:08:15.112 "superblock": true, 00:08:15.112 "num_base_bdevs": 2, 00:08:15.112 "num_base_bdevs_discovered": 0, 00:08:15.112 "num_base_bdevs_operational": 2, 00:08:15.112 "base_bdevs_list": [ 00:08:15.112 { 00:08:15.112 "name": "BaseBdev1", 00:08:15.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.112 "is_configured": false, 00:08:15.112 "data_offset": 0, 00:08:15.112 "data_size": 0 00:08:15.112 }, 00:08:15.112 { 00:08:15.112 "name": "BaseBdev2", 00:08:15.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.112 "is_configured": false, 00:08:15.112 "data_offset": 0, 00:08:15.112 "data_size": 0 00:08:15.112 } 00:08:15.112 ] 00:08:15.112 }' 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.112 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 [2024-10-29 10:57:20.900330] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:15.682 [2024-10-29 10:57:20.900460] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 [2024-10-29 10:57:20.912312] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:15.682 [2024-10-29 10:57:20.912402] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:15.682 [2024-10-29 10:57:20.912448] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:15.682 [2024-10-29 10:57:20.912471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 [2024-10-29 10:57:20.933073] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:15.682 BaseBdev1 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 [ 00:08:15.682 { 00:08:15.682 "name": "BaseBdev1", 00:08:15.682 "aliases": [ 00:08:15.682 "205d26ee-1d8e-4933-970b-12eb2ddbfb7e" 00:08:15.682 ], 00:08:15.682 "product_name": "Malloc disk", 00:08:15.682 "block_size": 512, 00:08:15.682 "num_blocks": 65536, 00:08:15.682 "uuid": "205d26ee-1d8e-4933-970b-12eb2ddbfb7e", 00:08:15.682 "assigned_rate_limits": { 00:08:15.682 "rw_ios_per_sec": 0, 00:08:15.682 "rw_mbytes_per_sec": 0, 00:08:15.682 "r_mbytes_per_sec": 0, 00:08:15.682 "w_mbytes_per_sec": 0 00:08:15.682 }, 00:08:15.682 "claimed": true, 00:08:15.682 "claim_type": "exclusive_write", 00:08:15.682 "zoned": false, 00:08:15.682 "supported_io_types": { 00:08:15.682 "read": true, 00:08:15.682 "write": true, 00:08:15.682 "unmap": true, 00:08:15.682 "flush": true, 00:08:15.682 "reset": true, 00:08:15.682 "nvme_admin": false, 00:08:15.682 "nvme_io": false, 00:08:15.682 "nvme_io_md": false, 00:08:15.682 "write_zeroes": true, 00:08:15.682 "zcopy": true, 00:08:15.682 "get_zone_info": false, 00:08:15.682 "zone_management": false, 00:08:15.682 "zone_append": false, 00:08:15.682 "compare": false, 00:08:15.682 "compare_and_write": false, 00:08:15.682 "abort": true, 00:08:15.682 "seek_hole": false, 00:08:15.682 "seek_data": false, 00:08:15.682 "copy": true, 00:08:15.682 "nvme_iov_md": false 00:08:15.682 }, 00:08:15.682 "memory_domains": [ 00:08:15.682 { 00:08:15.682 "dma_device_id": "system", 00:08:15.682 "dma_device_type": 1 00:08:15.682 }, 00:08:15.682 { 00:08:15.682 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:15.682 "dma_device_type": 2 00:08:15.682 } 00:08:15.682 ], 00:08:15.682 "driver_specific": {} 00:08:15.682 } 00:08:15.682 ] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.682 10:57:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.683 10:57:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.683 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.683 "name": "Existed_Raid", 00:08:15.683 "uuid": "a1df43be-2483-4a16-b6b3-6f18876dd852", 00:08:15.683 "strip_size_kb": 0, 00:08:15.683 "state": "configuring", 00:08:15.683 "raid_level": "raid1", 00:08:15.683 "superblock": true, 00:08:15.683 "num_base_bdevs": 2, 00:08:15.683 "num_base_bdevs_discovered": 1, 00:08:15.683 "num_base_bdevs_operational": 2, 00:08:15.683 "base_bdevs_list": [ 00:08:15.683 { 00:08:15.683 "name": "BaseBdev1", 00:08:15.683 "uuid": "205d26ee-1d8e-4933-970b-12eb2ddbfb7e", 00:08:15.683 "is_configured": true, 00:08:15.683 "data_offset": 2048, 00:08:15.683 "data_size": 63488 00:08:15.683 }, 00:08:15.683 { 00:08:15.683 "name": "BaseBdev2", 00:08:15.683 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.683 "is_configured": false, 00:08:15.683 "data_offset": 0, 00:08:15.683 "data_size": 0 00:08:15.683 } 00:08:15.683 ] 00:08:15.683 }' 00:08:15.683 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.683 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.943 [2024-10-29 10:57:21.392323] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:15.943 [2024-10-29 10:57:21.392452] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.943 [2024-10-29 10:57:21.404324] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:15.943 [2024-10-29 10:57:21.406249] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:15.943 [2024-10-29 10:57:21.406290] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.943 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.203 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.203 "name": "Existed_Raid", 00:08:16.203 "uuid": "9586d77a-8405-44cf-a120-f51fd6b67437", 00:08:16.203 "strip_size_kb": 0, 00:08:16.203 "state": "configuring", 00:08:16.203 "raid_level": "raid1", 00:08:16.203 "superblock": true, 00:08:16.203 "num_base_bdevs": 2, 00:08:16.203 "num_base_bdevs_discovered": 1, 00:08:16.203 "num_base_bdevs_operational": 2, 00:08:16.203 "base_bdevs_list": [ 00:08:16.203 { 00:08:16.203 "name": "BaseBdev1", 00:08:16.204 "uuid": "205d26ee-1d8e-4933-970b-12eb2ddbfb7e", 00:08:16.204 "is_configured": true, 00:08:16.204 "data_offset": 2048, 00:08:16.204 "data_size": 63488 00:08:16.204 }, 00:08:16.204 { 00:08:16.204 "name": "BaseBdev2", 00:08:16.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.204 "is_configured": false, 00:08:16.204 "data_offset": 0, 00:08:16.204 "data_size": 0 00:08:16.204 } 00:08:16.204 ] 00:08:16.204 }' 00:08:16.204 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.204 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:16.464 [2024-10-29 10:57:21.854408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:16.464 [2024-10-29 10:57:21.854771] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:16.464 [2024-10-29 10:57:21.854826] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:16.464 BaseBdev2 00:08:16.464 [2024-10-29 10:57:21.855164] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:16.464 [2024-10-29 10:57:21.855338] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:16.464 [2024-10-29 10:57:21.855426] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:16.464 [2024-10-29 10:57:21.855621] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:16.464 [ 00:08:16.464 { 00:08:16.464 "name": "BaseBdev2", 00:08:16.464 "aliases": [ 00:08:16.464 "0b381293-dca7-4838-b285-09b74a4bc128" 00:08:16.464 ], 00:08:16.464 "product_name": "Malloc disk", 00:08:16.464 "block_size": 512, 00:08:16.464 "num_blocks": 65536, 00:08:16.464 "uuid": "0b381293-dca7-4838-b285-09b74a4bc128", 00:08:16.464 "assigned_rate_limits": { 00:08:16.464 "rw_ios_per_sec": 0, 00:08:16.464 "rw_mbytes_per_sec": 0, 00:08:16.464 "r_mbytes_per_sec": 0, 00:08:16.464 "w_mbytes_per_sec": 0 00:08:16.464 }, 00:08:16.464 "claimed": true, 00:08:16.464 "claim_type": "exclusive_write", 00:08:16.464 "zoned": false, 00:08:16.464 "supported_io_types": { 00:08:16.464 "read": true, 00:08:16.464 "write": true, 00:08:16.464 "unmap": true, 00:08:16.464 "flush": true, 00:08:16.464 "reset": true, 00:08:16.464 "nvme_admin": false, 00:08:16.464 "nvme_io": false, 00:08:16.464 "nvme_io_md": false, 00:08:16.464 "write_zeroes": true, 00:08:16.464 "zcopy": true, 00:08:16.464 "get_zone_info": false, 00:08:16.464 "zone_management": false, 00:08:16.464 "zone_append": false, 00:08:16.464 "compare": false, 00:08:16.464 "compare_and_write": false, 00:08:16.464 "abort": true, 00:08:16.464 "seek_hole": false, 00:08:16.464 "seek_data": false, 00:08:16.464 "copy": true, 00:08:16.464 "nvme_iov_md": false 00:08:16.464 }, 00:08:16.464 "memory_domains": [ 00:08:16.464 { 00:08:16.464 "dma_device_id": "system", 00:08:16.464 "dma_device_type": 1 00:08:16.464 }, 00:08:16.464 { 00:08:16.464 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:16.464 "dma_device_type": 2 00:08:16.464 } 00:08:16.464 ], 00:08:16.464 "driver_specific": {} 00:08:16.464 } 00:08:16.464 ] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.464 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.464 "name": "Existed_Raid", 00:08:16.464 "uuid": "9586d77a-8405-44cf-a120-f51fd6b67437", 00:08:16.464 "strip_size_kb": 0, 00:08:16.464 "state": "online", 00:08:16.465 "raid_level": "raid1", 00:08:16.465 "superblock": true, 00:08:16.465 "num_base_bdevs": 2, 00:08:16.465 "num_base_bdevs_discovered": 2, 00:08:16.465 "num_base_bdevs_operational": 2, 00:08:16.465 "base_bdevs_list": [ 00:08:16.465 { 00:08:16.465 "name": "BaseBdev1", 00:08:16.465 "uuid": "205d26ee-1d8e-4933-970b-12eb2ddbfb7e", 00:08:16.465 "is_configured": true, 00:08:16.465 "data_offset": 2048, 00:08:16.465 "data_size": 63488 00:08:16.465 }, 00:08:16.465 { 00:08:16.465 "name": "BaseBdev2", 00:08:16.465 "uuid": "0b381293-dca7-4838-b285-09b74a4bc128", 00:08:16.465 "is_configured": true, 00:08:16.465 "data_offset": 2048, 00:08:16.465 "data_size": 63488 00:08:16.465 } 00:08:16.465 ] 00:08:16.465 }' 00:08:16.465 10:57:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.465 10:57:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.035 [2024-10-29 10:57:22.361869] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:17.035 "name": "Existed_Raid", 00:08:17.035 "aliases": [ 00:08:17.035 "9586d77a-8405-44cf-a120-f51fd6b67437" 00:08:17.035 ], 00:08:17.035 "product_name": "Raid Volume", 00:08:17.035 "block_size": 512, 00:08:17.035 "num_blocks": 63488, 00:08:17.035 "uuid": "9586d77a-8405-44cf-a120-f51fd6b67437", 00:08:17.035 "assigned_rate_limits": { 00:08:17.035 "rw_ios_per_sec": 0, 00:08:17.035 "rw_mbytes_per_sec": 0, 00:08:17.035 "r_mbytes_per_sec": 0, 00:08:17.035 "w_mbytes_per_sec": 0 00:08:17.035 }, 00:08:17.035 "claimed": false, 00:08:17.035 "zoned": false, 00:08:17.035 "supported_io_types": { 00:08:17.035 "read": true, 00:08:17.035 "write": true, 00:08:17.035 "unmap": false, 00:08:17.035 "flush": false, 00:08:17.035 "reset": true, 00:08:17.035 "nvme_admin": false, 00:08:17.035 "nvme_io": false, 00:08:17.035 "nvme_io_md": false, 00:08:17.035 "write_zeroes": true, 00:08:17.035 "zcopy": false, 00:08:17.035 "get_zone_info": false, 00:08:17.035 "zone_management": false, 00:08:17.035 "zone_append": false, 00:08:17.035 "compare": false, 00:08:17.035 "compare_and_write": false, 00:08:17.035 "abort": false, 00:08:17.035 "seek_hole": false, 00:08:17.035 "seek_data": false, 00:08:17.035 "copy": false, 00:08:17.035 "nvme_iov_md": false 00:08:17.035 }, 00:08:17.035 "memory_domains": [ 00:08:17.035 { 00:08:17.035 "dma_device_id": "system", 00:08:17.035 "dma_device_type": 1 00:08:17.035 }, 00:08:17.035 { 00:08:17.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:17.035 "dma_device_type": 2 00:08:17.035 }, 00:08:17.035 { 00:08:17.035 "dma_device_id": "system", 00:08:17.035 "dma_device_type": 1 00:08:17.035 }, 00:08:17.035 { 00:08:17.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:17.035 "dma_device_type": 2 00:08:17.035 } 00:08:17.035 ], 00:08:17.035 "driver_specific": { 00:08:17.035 "raid": { 00:08:17.035 "uuid": "9586d77a-8405-44cf-a120-f51fd6b67437", 00:08:17.035 "strip_size_kb": 0, 00:08:17.035 "state": "online", 00:08:17.035 "raid_level": "raid1", 00:08:17.035 "superblock": true, 00:08:17.035 "num_base_bdevs": 2, 00:08:17.035 "num_base_bdevs_discovered": 2, 00:08:17.035 "num_base_bdevs_operational": 2, 00:08:17.035 "base_bdevs_list": [ 00:08:17.035 { 00:08:17.035 "name": "BaseBdev1", 00:08:17.035 "uuid": "205d26ee-1d8e-4933-970b-12eb2ddbfb7e", 00:08:17.035 "is_configured": true, 00:08:17.035 "data_offset": 2048, 00:08:17.035 "data_size": 63488 00:08:17.035 }, 00:08:17.035 { 00:08:17.035 "name": "BaseBdev2", 00:08:17.035 "uuid": "0b381293-dca7-4838-b285-09b74a4bc128", 00:08:17.035 "is_configured": true, 00:08:17.035 "data_offset": 2048, 00:08:17.035 "data_size": 63488 00:08:17.035 } 00:08:17.035 ] 00:08:17.035 } 00:08:17.035 } 00:08:17.035 }' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:17.035 BaseBdev2' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:17.035 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.036 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.296 [2024-10-29 10:57:22.573299] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.296 "name": "Existed_Raid", 00:08:17.296 "uuid": "9586d77a-8405-44cf-a120-f51fd6b67437", 00:08:17.296 "strip_size_kb": 0, 00:08:17.296 "state": "online", 00:08:17.296 "raid_level": "raid1", 00:08:17.296 "superblock": true, 00:08:17.296 "num_base_bdevs": 2, 00:08:17.296 "num_base_bdevs_discovered": 1, 00:08:17.296 "num_base_bdevs_operational": 1, 00:08:17.296 "base_bdevs_list": [ 00:08:17.296 { 00:08:17.296 "name": null, 00:08:17.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.296 "is_configured": false, 00:08:17.296 "data_offset": 0, 00:08:17.296 "data_size": 63488 00:08:17.296 }, 00:08:17.296 { 00:08:17.296 "name": "BaseBdev2", 00:08:17.296 "uuid": "0b381293-dca7-4838-b285-09b74a4bc128", 00:08:17.296 "is_configured": true, 00:08:17.296 "data_offset": 2048, 00:08:17.296 "data_size": 63488 00:08:17.296 } 00:08:17.296 ] 00:08:17.296 }' 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.296 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.556 10:57:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.556 [2024-10-29 10:57:23.019633] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:17.556 [2024-10-29 10:57:23.019796] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:17.556 [2024-10-29 10:57:23.031306] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.556 [2024-10-29 10:57:23.031446] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.556 [2024-10-29 10:57:23.031464] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:17.556 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74365 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 74365 ']' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 74365 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 74365 00:08:17.816 killing process with pid 74365 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 74365' 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 74365 00:08:17.816 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 74365 00:08:17.816 [2024-10-29 10:57:23.130647] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:17.816 [2024-10-29 10:57:23.131721] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:18.076 10:57:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:18.076 00:08:18.076 real 0m3.765s 00:08:18.076 user 0m5.924s 00:08:18.076 sys 0m0.781s 00:08:18.076 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:18.076 10:57:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:18.076 ************************************ 00:08:18.076 END TEST raid_state_function_test_sb 00:08:18.076 ************************************ 00:08:18.076 10:57:23 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:08:18.076 10:57:23 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:08:18.076 10:57:23 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:18.076 10:57:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:18.076 ************************************ 00:08:18.076 START TEST raid_superblock_test 00:08:18.076 ************************************ 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 2 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:08:18.076 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74605 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74605 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 74605 ']' 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.077 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:18.077 10:57:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.077 [2024-10-29 10:57:23.500402] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:18.077 [2024-10-29 10:57:23.500527] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74605 ] 00:08:18.336 [2024-10-29 10:57:23.671869] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.336 [2024-10-29 10:57:23.697290] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.336 [2024-10-29 10:57:23.738711] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.336 [2024-10-29 10:57:23.738833] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.907 malloc1 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.907 [2024-10-29 10:57:24.361008] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:18.907 [2024-10-29 10:57:24.361140] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:18.907 [2024-10-29 10:57:24.361179] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:18.907 [2024-10-29 10:57:24.361222] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:18.907 [2024-10-29 10:57:24.363333] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:18.907 [2024-10-29 10:57:24.363424] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:18.907 pt1 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.907 malloc2 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.907 [2024-10-29 10:57:24.393789] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:18.907 [2024-10-29 10:57:24.393858] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:18.907 [2024-10-29 10:57:24.393878] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:18.907 [2024-10-29 10:57:24.393889] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:18.907 [2024-10-29 10:57:24.396067] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:18.907 [2024-10-29 10:57:24.396168] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:18.907 pt2 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.907 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.167 [2024-10-29 10:57:24.405838] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:19.167 [2024-10-29 10:57:24.407793] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:19.167 [2024-10-29 10:57:24.407952] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:08:19.167 [2024-10-29 10:57:24.407968] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:19.167 [2024-10-29 10:57:24.408270] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:19.167 [2024-10-29 10:57:24.408466] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:08:19.167 [2024-10-29 10:57:24.408477] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:08:19.167 [2024-10-29 10:57:24.408625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.167 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.168 "name": "raid_bdev1", 00:08:19.168 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:19.168 "strip_size_kb": 0, 00:08:19.168 "state": "online", 00:08:19.168 "raid_level": "raid1", 00:08:19.168 "superblock": true, 00:08:19.168 "num_base_bdevs": 2, 00:08:19.168 "num_base_bdevs_discovered": 2, 00:08:19.168 "num_base_bdevs_operational": 2, 00:08:19.168 "base_bdevs_list": [ 00:08:19.168 { 00:08:19.168 "name": "pt1", 00:08:19.168 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:19.168 "is_configured": true, 00:08:19.168 "data_offset": 2048, 00:08:19.168 "data_size": 63488 00:08:19.168 }, 00:08:19.168 { 00:08:19.168 "name": "pt2", 00:08:19.168 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:19.168 "is_configured": true, 00:08:19.168 "data_offset": 2048, 00:08:19.168 "data_size": 63488 00:08:19.168 } 00:08:19.168 ] 00:08:19.168 }' 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.168 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.428 [2024-10-29 10:57:24.841297] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:19.428 "name": "raid_bdev1", 00:08:19.428 "aliases": [ 00:08:19.428 "fe19b567-8361-495a-bbe4-7fee815ca4a3" 00:08:19.428 ], 00:08:19.428 "product_name": "Raid Volume", 00:08:19.428 "block_size": 512, 00:08:19.428 "num_blocks": 63488, 00:08:19.428 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:19.428 "assigned_rate_limits": { 00:08:19.428 "rw_ios_per_sec": 0, 00:08:19.428 "rw_mbytes_per_sec": 0, 00:08:19.428 "r_mbytes_per_sec": 0, 00:08:19.428 "w_mbytes_per_sec": 0 00:08:19.428 }, 00:08:19.428 "claimed": false, 00:08:19.428 "zoned": false, 00:08:19.428 "supported_io_types": { 00:08:19.428 "read": true, 00:08:19.428 "write": true, 00:08:19.428 "unmap": false, 00:08:19.428 "flush": false, 00:08:19.428 "reset": true, 00:08:19.428 "nvme_admin": false, 00:08:19.428 "nvme_io": false, 00:08:19.428 "nvme_io_md": false, 00:08:19.428 "write_zeroes": true, 00:08:19.428 "zcopy": false, 00:08:19.428 "get_zone_info": false, 00:08:19.428 "zone_management": false, 00:08:19.428 "zone_append": false, 00:08:19.428 "compare": false, 00:08:19.428 "compare_and_write": false, 00:08:19.428 "abort": false, 00:08:19.428 "seek_hole": false, 00:08:19.428 "seek_data": false, 00:08:19.428 "copy": false, 00:08:19.428 "nvme_iov_md": false 00:08:19.428 }, 00:08:19.428 "memory_domains": [ 00:08:19.428 { 00:08:19.428 "dma_device_id": "system", 00:08:19.428 "dma_device_type": 1 00:08:19.428 }, 00:08:19.428 { 00:08:19.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.428 "dma_device_type": 2 00:08:19.428 }, 00:08:19.428 { 00:08:19.428 "dma_device_id": "system", 00:08:19.428 "dma_device_type": 1 00:08:19.428 }, 00:08:19.428 { 00:08:19.428 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.428 "dma_device_type": 2 00:08:19.428 } 00:08:19.428 ], 00:08:19.428 "driver_specific": { 00:08:19.428 "raid": { 00:08:19.428 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:19.428 "strip_size_kb": 0, 00:08:19.428 "state": "online", 00:08:19.428 "raid_level": "raid1", 00:08:19.428 "superblock": true, 00:08:19.428 "num_base_bdevs": 2, 00:08:19.428 "num_base_bdevs_discovered": 2, 00:08:19.428 "num_base_bdevs_operational": 2, 00:08:19.428 "base_bdevs_list": [ 00:08:19.428 { 00:08:19.428 "name": "pt1", 00:08:19.428 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:19.428 "is_configured": true, 00:08:19.428 "data_offset": 2048, 00:08:19.428 "data_size": 63488 00:08:19.428 }, 00:08:19.428 { 00:08:19.428 "name": "pt2", 00:08:19.428 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:19.428 "is_configured": true, 00:08:19.428 "data_offset": 2048, 00:08:19.428 "data_size": 63488 00:08:19.428 } 00:08:19.428 ] 00:08:19.428 } 00:08:19.428 } 00:08:19.428 }' 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:19.428 pt2' 00:08:19.428 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.689 10:57:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:19.689 [2024-10-29 10:57:25.088774] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=fe19b567-8361-495a-bbe4-7fee815ca4a3 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z fe19b567-8361-495a-bbe4-7fee815ca4a3 ']' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.689 [2024-10-29 10:57:25.132470] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:19.689 [2024-10-29 10:57:25.132494] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:19.689 [2024-10-29 10:57:25.132568] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:19.689 [2024-10-29 10:57:25.132630] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:19.689 [2024-10-29 10:57:25.132639] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.689 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 [2024-10-29 10:57:25.264330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:19.949 [2024-10-29 10:57:25.266256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:19.949 [2024-10-29 10:57:25.266333] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:19.949 [2024-10-29 10:57:25.266386] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:19.949 [2024-10-29 10:57:25.266404] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:19.949 [2024-10-29 10:57:25.266413] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:19.949 request: 00:08:19.949 { 00:08:19.949 "name": "raid_bdev1", 00:08:19.949 "raid_level": "raid1", 00:08:19.949 "base_bdevs": [ 00:08:19.949 "malloc1", 00:08:19.949 "malloc2" 00:08:19.949 ], 00:08:19.949 "superblock": false, 00:08:19.949 "method": "bdev_raid_create", 00:08:19.949 "req_id": 1 00:08:19.949 } 00:08:19.949 Got JSON-RPC error response 00:08:19.949 response: 00:08:19.949 { 00:08:19.949 "code": -17, 00:08:19.949 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:19.949 } 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.949 [2024-10-29 10:57:25.316185] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:19.949 [2024-10-29 10:57:25.316299] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.949 [2024-10-29 10:57:25.316337] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:19.949 [2024-10-29 10:57:25.316382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.949 [2024-10-29 10:57:25.318674] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.949 [2024-10-29 10:57:25.318740] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:19.949 [2024-10-29 10:57:25.318862] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:19.949 [2024-10-29 10:57:25.318933] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:19.949 pt1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:19.949 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.950 "name": "raid_bdev1", 00:08:19.950 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:19.950 "strip_size_kb": 0, 00:08:19.950 "state": "configuring", 00:08:19.950 "raid_level": "raid1", 00:08:19.950 "superblock": true, 00:08:19.950 "num_base_bdevs": 2, 00:08:19.950 "num_base_bdevs_discovered": 1, 00:08:19.950 "num_base_bdevs_operational": 2, 00:08:19.950 "base_bdevs_list": [ 00:08:19.950 { 00:08:19.950 "name": "pt1", 00:08:19.950 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:19.950 "is_configured": true, 00:08:19.950 "data_offset": 2048, 00:08:19.950 "data_size": 63488 00:08:19.950 }, 00:08:19.950 { 00:08:19.950 "name": null, 00:08:19.950 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:19.950 "is_configured": false, 00:08:19.950 "data_offset": 2048, 00:08:19.950 "data_size": 63488 00:08:19.950 } 00:08:19.950 ] 00:08:19.950 }' 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.950 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.519 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.519 [2024-10-29 10:57:25.759443] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:20.519 [2024-10-29 10:57:25.759513] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:20.519 [2024-10-29 10:57:25.759537] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:20.519 [2024-10-29 10:57:25.759546] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:20.519 [2024-10-29 10:57:25.759954] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:20.519 [2024-10-29 10:57:25.759969] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:20.519 [2024-10-29 10:57:25.760042] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:20.520 [2024-10-29 10:57:25.760067] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:20.520 [2024-10-29 10:57:25.760168] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:20.520 [2024-10-29 10:57:25.760177] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:20.520 [2024-10-29 10:57:25.760401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:20.520 [2024-10-29 10:57:25.760516] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:20.520 [2024-10-29 10:57:25.760530] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:20.520 [2024-10-29 10:57:25.760650] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:20.520 pt2 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.520 "name": "raid_bdev1", 00:08:20.520 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:20.520 "strip_size_kb": 0, 00:08:20.520 "state": "online", 00:08:20.520 "raid_level": "raid1", 00:08:20.520 "superblock": true, 00:08:20.520 "num_base_bdevs": 2, 00:08:20.520 "num_base_bdevs_discovered": 2, 00:08:20.520 "num_base_bdevs_operational": 2, 00:08:20.520 "base_bdevs_list": [ 00:08:20.520 { 00:08:20.520 "name": "pt1", 00:08:20.520 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:20.520 "is_configured": true, 00:08:20.520 "data_offset": 2048, 00:08:20.520 "data_size": 63488 00:08:20.520 }, 00:08:20.520 { 00:08:20.520 "name": "pt2", 00:08:20.520 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:20.520 "is_configured": true, 00:08:20.520 "data_offset": 2048, 00:08:20.520 "data_size": 63488 00:08:20.520 } 00:08:20.520 ] 00:08:20.520 }' 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.520 10:57:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.781 [2024-10-29 10:57:26.230910] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:20.781 "name": "raid_bdev1", 00:08:20.781 "aliases": [ 00:08:20.781 "fe19b567-8361-495a-bbe4-7fee815ca4a3" 00:08:20.781 ], 00:08:20.781 "product_name": "Raid Volume", 00:08:20.781 "block_size": 512, 00:08:20.781 "num_blocks": 63488, 00:08:20.781 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:20.781 "assigned_rate_limits": { 00:08:20.781 "rw_ios_per_sec": 0, 00:08:20.781 "rw_mbytes_per_sec": 0, 00:08:20.781 "r_mbytes_per_sec": 0, 00:08:20.781 "w_mbytes_per_sec": 0 00:08:20.781 }, 00:08:20.781 "claimed": false, 00:08:20.781 "zoned": false, 00:08:20.781 "supported_io_types": { 00:08:20.781 "read": true, 00:08:20.781 "write": true, 00:08:20.781 "unmap": false, 00:08:20.781 "flush": false, 00:08:20.781 "reset": true, 00:08:20.781 "nvme_admin": false, 00:08:20.781 "nvme_io": false, 00:08:20.781 "nvme_io_md": false, 00:08:20.781 "write_zeroes": true, 00:08:20.781 "zcopy": false, 00:08:20.781 "get_zone_info": false, 00:08:20.781 "zone_management": false, 00:08:20.781 "zone_append": false, 00:08:20.781 "compare": false, 00:08:20.781 "compare_and_write": false, 00:08:20.781 "abort": false, 00:08:20.781 "seek_hole": false, 00:08:20.781 "seek_data": false, 00:08:20.781 "copy": false, 00:08:20.781 "nvme_iov_md": false 00:08:20.781 }, 00:08:20.781 "memory_domains": [ 00:08:20.781 { 00:08:20.781 "dma_device_id": "system", 00:08:20.781 "dma_device_type": 1 00:08:20.781 }, 00:08:20.781 { 00:08:20.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.781 "dma_device_type": 2 00:08:20.781 }, 00:08:20.781 { 00:08:20.781 "dma_device_id": "system", 00:08:20.781 "dma_device_type": 1 00:08:20.781 }, 00:08:20.781 { 00:08:20.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:20.781 "dma_device_type": 2 00:08:20.781 } 00:08:20.781 ], 00:08:20.781 "driver_specific": { 00:08:20.781 "raid": { 00:08:20.781 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:20.781 "strip_size_kb": 0, 00:08:20.781 "state": "online", 00:08:20.781 "raid_level": "raid1", 00:08:20.781 "superblock": true, 00:08:20.781 "num_base_bdevs": 2, 00:08:20.781 "num_base_bdevs_discovered": 2, 00:08:20.781 "num_base_bdevs_operational": 2, 00:08:20.781 "base_bdevs_list": [ 00:08:20.781 { 00:08:20.781 "name": "pt1", 00:08:20.781 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:20.781 "is_configured": true, 00:08:20.781 "data_offset": 2048, 00:08:20.781 "data_size": 63488 00:08:20.781 }, 00:08:20.781 { 00:08:20.781 "name": "pt2", 00:08:20.781 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:20.781 "is_configured": true, 00:08:20.781 "data_offset": 2048, 00:08:20.781 "data_size": 63488 00:08:20.781 } 00:08:20.781 ] 00:08:20.781 } 00:08:20.781 } 00:08:20.781 }' 00:08:20.781 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:21.042 pt2' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.042 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:21.043 [2024-10-29 10:57:26.410627] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' fe19b567-8361-495a-bbe4-7fee815ca4a3 '!=' fe19b567-8361-495a-bbe4-7fee815ca4a3 ']' 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.043 [2024-10-29 10:57:26.442562] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.043 "name": "raid_bdev1", 00:08:21.043 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:21.043 "strip_size_kb": 0, 00:08:21.043 "state": "online", 00:08:21.043 "raid_level": "raid1", 00:08:21.043 "superblock": true, 00:08:21.043 "num_base_bdevs": 2, 00:08:21.043 "num_base_bdevs_discovered": 1, 00:08:21.043 "num_base_bdevs_operational": 1, 00:08:21.043 "base_bdevs_list": [ 00:08:21.043 { 00:08:21.043 "name": null, 00:08:21.043 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.043 "is_configured": false, 00:08:21.043 "data_offset": 0, 00:08:21.043 "data_size": 63488 00:08:21.043 }, 00:08:21.043 { 00:08:21.043 "name": "pt2", 00:08:21.043 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:21.043 "is_configured": true, 00:08:21.043 "data_offset": 2048, 00:08:21.043 "data_size": 63488 00:08:21.043 } 00:08:21.043 ] 00:08:21.043 }' 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.043 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.613 [2024-10-29 10:57:26.889522] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:21.613 [2024-10-29 10:57:26.889575] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:21.613 [2024-10-29 10:57:26.889679] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:21.613 [2024-10-29 10:57:26.889743] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:21.613 [2024-10-29 10:57:26.889763] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:21.613 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.614 [2024-10-29 10:57:26.961352] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:21.614 [2024-10-29 10:57:26.961423] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:21.614 [2024-10-29 10:57:26.961445] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:21.614 [2024-10-29 10:57:26.961455] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:21.614 [2024-10-29 10:57:26.964117] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:21.614 [2024-10-29 10:57:26.964151] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:21.614 [2024-10-29 10:57:26.964242] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:21.614 [2024-10-29 10:57:26.964281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:21.614 [2024-10-29 10:57:26.964398] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:21.614 [2024-10-29 10:57:26.964408] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:21.614 [2024-10-29 10:57:26.964655] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:21.614 [2024-10-29 10:57:26.964809] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:21.614 [2024-10-29 10:57:26.964829] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:21.614 [2024-10-29 10:57:26.964960] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:21.614 pt2 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.614 10:57:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.614 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.614 "name": "raid_bdev1", 00:08:21.614 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:21.614 "strip_size_kb": 0, 00:08:21.614 "state": "online", 00:08:21.614 "raid_level": "raid1", 00:08:21.614 "superblock": true, 00:08:21.614 "num_base_bdevs": 2, 00:08:21.614 "num_base_bdevs_discovered": 1, 00:08:21.614 "num_base_bdevs_operational": 1, 00:08:21.614 "base_bdevs_list": [ 00:08:21.614 { 00:08:21.614 "name": null, 00:08:21.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.614 "is_configured": false, 00:08:21.614 "data_offset": 2048, 00:08:21.614 "data_size": 63488 00:08:21.614 }, 00:08:21.614 { 00:08:21.614 "name": "pt2", 00:08:21.614 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:21.614 "is_configured": true, 00:08:21.614 "data_offset": 2048, 00:08:21.614 "data_size": 63488 00:08:21.614 } 00:08:21.614 ] 00:08:21.614 }' 00:08:21.614 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.614 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.184 [2024-10-29 10:57:27.396694] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:22.184 [2024-10-29 10:57:27.396743] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:22.184 [2024-10-29 10:57:27.396844] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:22.184 [2024-10-29 10:57:27.396909] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:22.184 [2024-10-29 10:57:27.396925] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.184 [2024-10-29 10:57:27.460520] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:22.184 [2024-10-29 10:57:27.460588] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:22.184 [2024-10-29 10:57:27.460614] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:08:22.184 [2024-10-29 10:57:27.460632] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:22.184 [2024-10-29 10:57:27.463283] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:22.184 [2024-10-29 10:57:27.463321] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:22.184 [2024-10-29 10:57:27.463421] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:22.184 [2024-10-29 10:57:27.463471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:22.184 [2024-10-29 10:57:27.463581] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:08:22.184 [2024-10-29 10:57:27.463593] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:22.184 [2024-10-29 10:57:27.463612] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:08:22.184 [2024-10-29 10:57:27.463663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:22.184 [2024-10-29 10:57:27.463741] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:08:22.184 [2024-10-29 10:57:27.463752] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:22.184 [2024-10-29 10:57:27.464023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:08:22.184 [2024-10-29 10:57:27.464167] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:08:22.184 [2024-10-29 10:57:27.464200] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:08:22.184 [2024-10-29 10:57:27.464336] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:22.184 pt1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.184 "name": "raid_bdev1", 00:08:22.184 "uuid": "fe19b567-8361-495a-bbe4-7fee815ca4a3", 00:08:22.184 "strip_size_kb": 0, 00:08:22.184 "state": "online", 00:08:22.184 "raid_level": "raid1", 00:08:22.184 "superblock": true, 00:08:22.184 "num_base_bdevs": 2, 00:08:22.184 "num_base_bdevs_discovered": 1, 00:08:22.184 "num_base_bdevs_operational": 1, 00:08:22.184 "base_bdevs_list": [ 00:08:22.184 { 00:08:22.184 "name": null, 00:08:22.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.184 "is_configured": false, 00:08:22.184 "data_offset": 2048, 00:08:22.184 "data_size": 63488 00:08:22.184 }, 00:08:22.184 { 00:08:22.184 "name": "pt2", 00:08:22.184 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:22.184 "is_configured": true, 00:08:22.184 "data_offset": 2048, 00:08:22.184 "data_size": 63488 00:08:22.184 } 00:08:22.184 ] 00:08:22.184 }' 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.184 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.444 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:08:22.444 [2024-10-29 10:57:27.936109] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' fe19b567-8361-495a-bbe4-7fee815ca4a3 '!=' fe19b567-8361-495a-bbe4-7fee815ca4a3 ']' 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74605 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 74605 ']' 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 74605 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:22.704 10:57:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 74605 00:08:22.704 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:22.704 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:22.704 killing process with pid 74605 00:08:22.704 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 74605' 00:08:22.704 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 74605 00:08:22.704 [2024-10-29 10:57:28.008711] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:22.704 [2024-10-29 10:57:28.008837] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:22.704 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 74605 00:08:22.704 [2024-10-29 10:57:28.008909] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:22.704 [2024-10-29 10:57:28.008920] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:08:22.704 [2024-10-29 10:57:28.051641] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:22.965 10:57:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:22.965 00:08:22.965 real 0m4.865s 00:08:22.965 user 0m7.934s 00:08:22.965 sys 0m0.987s 00:08:22.965 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:22.965 10:57:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.965 ************************************ 00:08:22.965 END TEST raid_superblock_test 00:08:22.965 ************************************ 00:08:22.965 10:57:28 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:08:22.965 10:57:28 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:22.965 10:57:28 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:22.965 10:57:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:22.965 ************************************ 00:08:22.965 START TEST raid_read_error_test 00:08:22.965 ************************************ 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 2 read 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Do1jxRUwvw 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74914 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74914 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 74914 ']' 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:22.965 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:22.965 10:57:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.965 [2024-10-29 10:57:28.445746] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:22.965 [2024-10-29 10:57:28.445872] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74914 ] 00:08:23.225 [2024-10-29 10:57:28.617090] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:23.225 [2024-10-29 10:57:28.643101] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:23.225 [2024-10-29 10:57:28.685421] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:23.225 [2024-10-29 10:57:28.685465] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.795 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 BaseBdev1_malloc 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 true 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 [2024-10-29 10:57:29.319798] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:24.055 [2024-10-29 10:57:29.319855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:24.055 [2024-10-29 10:57:29.319878] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:24.055 [2024-10-29 10:57:29.319887] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:24.055 [2024-10-29 10:57:29.322184] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:24.055 [2024-10-29 10:57:29.322218] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:24.055 BaseBdev1 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 BaseBdev2_malloc 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 true 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 [2024-10-29 10:57:29.360642] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:24.055 [2024-10-29 10:57:29.360693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:24.055 [2024-10-29 10:57:29.360714] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:24.055 [2024-10-29 10:57:29.360722] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:24.055 [2024-10-29 10:57:29.362814] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:24.055 [2024-10-29 10:57:29.362845] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:24.055 BaseBdev2 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 [2024-10-29 10:57:29.372663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:24.055 [2024-10-29 10:57:29.374520] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:24.055 [2024-10-29 10:57:29.374705] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:24.055 [2024-10-29 10:57:29.374718] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:24.055 [2024-10-29 10:57:29.374967] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:24.055 [2024-10-29 10:57:29.375161] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:24.055 [2024-10-29 10:57:29.375182] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:24.055 [2024-10-29 10:57:29.375313] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.055 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.055 "name": "raid_bdev1", 00:08:24.055 "uuid": "e4c45277-d5c3-4005-9495-dabe5f7b1f36", 00:08:24.055 "strip_size_kb": 0, 00:08:24.055 "state": "online", 00:08:24.055 "raid_level": "raid1", 00:08:24.055 "superblock": true, 00:08:24.055 "num_base_bdevs": 2, 00:08:24.055 "num_base_bdevs_discovered": 2, 00:08:24.055 "num_base_bdevs_operational": 2, 00:08:24.055 "base_bdevs_list": [ 00:08:24.055 { 00:08:24.055 "name": "BaseBdev1", 00:08:24.055 "uuid": "e7a77cf0-03e0-569e-8681-35421f1d9dc3", 00:08:24.055 "is_configured": true, 00:08:24.055 "data_offset": 2048, 00:08:24.055 "data_size": 63488 00:08:24.055 }, 00:08:24.055 { 00:08:24.055 "name": "BaseBdev2", 00:08:24.055 "uuid": "7591b83e-d827-52ca-976b-5325ae58510f", 00:08:24.055 "is_configured": true, 00:08:24.055 "data_offset": 2048, 00:08:24.055 "data_size": 63488 00:08:24.055 } 00:08:24.055 ] 00:08:24.055 }' 00:08:24.056 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.056 10:57:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.319 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:24.319 10:57:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:24.577 [2024-10-29 10:57:29.896194] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.514 "name": "raid_bdev1", 00:08:25.514 "uuid": "e4c45277-d5c3-4005-9495-dabe5f7b1f36", 00:08:25.514 "strip_size_kb": 0, 00:08:25.514 "state": "online", 00:08:25.514 "raid_level": "raid1", 00:08:25.514 "superblock": true, 00:08:25.514 "num_base_bdevs": 2, 00:08:25.514 "num_base_bdevs_discovered": 2, 00:08:25.514 "num_base_bdevs_operational": 2, 00:08:25.514 "base_bdevs_list": [ 00:08:25.514 { 00:08:25.514 "name": "BaseBdev1", 00:08:25.514 "uuid": "e7a77cf0-03e0-569e-8681-35421f1d9dc3", 00:08:25.514 "is_configured": true, 00:08:25.514 "data_offset": 2048, 00:08:25.514 "data_size": 63488 00:08:25.514 }, 00:08:25.514 { 00:08:25.514 "name": "BaseBdev2", 00:08:25.514 "uuid": "7591b83e-d827-52ca-976b-5325ae58510f", 00:08:25.514 "is_configured": true, 00:08:25.514 "data_offset": 2048, 00:08:25.514 "data_size": 63488 00:08:25.514 } 00:08:25.514 ] 00:08:25.514 }' 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.514 10:57:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.775 [2024-10-29 10:57:31.243283] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:25.775 [2024-10-29 10:57:31.243318] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:25.775 [2024-10-29 10:57:31.245816] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:25.775 [2024-10-29 10:57:31.245861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:25.775 [2024-10-29 10:57:31.245942] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:25.775 [2024-10-29 10:57:31.245952] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:25.775 { 00:08:25.775 "results": [ 00:08:25.775 { 00:08:25.775 "job": "raid_bdev1", 00:08:25.775 "core_mask": "0x1", 00:08:25.775 "workload": "randrw", 00:08:25.775 "percentage": 50, 00:08:25.775 "status": "finished", 00:08:25.775 "queue_depth": 1, 00:08:25.775 "io_size": 131072, 00:08:25.775 "runtime": 1.347875, 00:08:25.775 "iops": 19373.458221274228, 00:08:25.775 "mibps": 2421.6822776592785, 00:08:25.775 "io_failed": 0, 00:08:25.775 "io_timeout": 0, 00:08:25.775 "avg_latency_us": 49.11724518748463, 00:08:25.775 "min_latency_us": 22.91703056768559, 00:08:25.775 "max_latency_us": 1380.8349344978167 00:08:25.775 } 00:08:25.775 ], 00:08:25.775 "core_count": 1 00:08:25.775 } 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74914 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 74914 ']' 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 74914 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:25.775 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 74914 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:26.035 killing process with pid 74914 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 74914' 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 74914 00:08:26.035 [2024-10-29 10:57:31.288119] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 74914 00:08:26.035 [2024-10-29 10:57:31.304180] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Do1jxRUwvw 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:26.035 00:08:26.035 real 0m3.166s 00:08:26.035 user 0m4.006s 00:08:26.035 sys 0m0.503s 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:26.035 10:57:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.035 ************************************ 00:08:26.035 END TEST raid_read_error_test 00:08:26.035 ************************************ 00:08:26.295 10:57:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:08:26.295 10:57:31 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:26.295 10:57:31 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:26.295 10:57:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:26.295 ************************************ 00:08:26.295 START TEST raid_write_error_test 00:08:26.295 ************************************ 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 2 write 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.CmreJGJYHz 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=75043 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 75043 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 75043 ']' 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:26.295 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:26.295 10:57:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.295 [2024-10-29 10:57:31.683609] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:26.295 [2024-10-29 10:57:31.683746] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75043 ] 00:08:26.556 [2024-10-29 10:57:31.854021] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:26.556 [2024-10-29 10:57:31.880550] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:26.556 [2024-10-29 10:57:31.923090] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:26.556 [2024-10-29 10:57:31.923146] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.125 BaseBdev1_malloc 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.125 true 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.125 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.126 [2024-10-29 10:57:32.545320] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:27.126 [2024-10-29 10:57:32.545385] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:27.126 [2024-10-29 10:57:32.545411] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:27.126 [2024-10-29 10:57:32.545422] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:27.126 [2024-10-29 10:57:32.547493] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:27.126 [2024-10-29 10:57:32.547527] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:27.126 BaseBdev1 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.126 BaseBdev2_malloc 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.126 true 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.126 [2024-10-29 10:57:32.585832] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:27.126 [2024-10-29 10:57:32.585878] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:27.126 [2024-10-29 10:57:32.585895] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:27.126 [2024-10-29 10:57:32.585904] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:27.126 [2024-10-29 10:57:32.587905] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:27.126 [2024-10-29 10:57:32.587938] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:27.126 BaseBdev2 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.126 [2024-10-29 10:57:32.597863] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:27.126 [2024-10-29 10:57:32.599654] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:27.126 [2024-10-29 10:57:32.599825] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:27.126 [2024-10-29 10:57:32.599844] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:27.126 [2024-10-29 10:57:32.600103] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:08:27.126 [2024-10-29 10:57:32.600262] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:27.126 [2024-10-29 10:57:32.600283] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:27.126 [2024-10-29 10:57:32.600413] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.126 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.386 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.386 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.386 "name": "raid_bdev1", 00:08:27.386 "uuid": "6bd2578d-61a2-4c94-93f6-abcefac1fefa", 00:08:27.386 "strip_size_kb": 0, 00:08:27.386 "state": "online", 00:08:27.386 "raid_level": "raid1", 00:08:27.386 "superblock": true, 00:08:27.386 "num_base_bdevs": 2, 00:08:27.386 "num_base_bdevs_discovered": 2, 00:08:27.386 "num_base_bdevs_operational": 2, 00:08:27.386 "base_bdevs_list": [ 00:08:27.386 { 00:08:27.386 "name": "BaseBdev1", 00:08:27.386 "uuid": "fb26f31b-9af6-56d9-9d4b-44ceba43277d", 00:08:27.386 "is_configured": true, 00:08:27.386 "data_offset": 2048, 00:08:27.386 "data_size": 63488 00:08:27.386 }, 00:08:27.386 { 00:08:27.386 "name": "BaseBdev2", 00:08:27.386 "uuid": "8c92aee2-0a52-51b0-a4b1-f66a5878ad95", 00:08:27.386 "is_configured": true, 00:08:27.386 "data_offset": 2048, 00:08:27.386 "data_size": 63488 00:08:27.386 } 00:08:27.386 ] 00:08:27.386 }' 00:08:27.386 10:57:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.386 10:57:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.647 10:57:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:27.647 10:57:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:27.647 [2024-10-29 10:57:33.125341] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.587 [2024-10-29 10:57:34.037964] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:08:28.587 [2024-10-29 10:57:34.038018] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:28.587 [2024-10-29 10:57:34.038231] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005d40 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.587 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.588 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.846 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.846 "name": "raid_bdev1", 00:08:28.846 "uuid": "6bd2578d-61a2-4c94-93f6-abcefac1fefa", 00:08:28.846 "strip_size_kb": 0, 00:08:28.846 "state": "online", 00:08:28.846 "raid_level": "raid1", 00:08:28.846 "superblock": true, 00:08:28.846 "num_base_bdevs": 2, 00:08:28.846 "num_base_bdevs_discovered": 1, 00:08:28.846 "num_base_bdevs_operational": 1, 00:08:28.846 "base_bdevs_list": [ 00:08:28.846 { 00:08:28.846 "name": null, 00:08:28.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.846 "is_configured": false, 00:08:28.846 "data_offset": 0, 00:08:28.846 "data_size": 63488 00:08:28.846 }, 00:08:28.846 { 00:08:28.846 "name": "BaseBdev2", 00:08:28.846 "uuid": "8c92aee2-0a52-51b0-a4b1-f66a5878ad95", 00:08:28.846 "is_configured": true, 00:08:28.846 "data_offset": 2048, 00:08:28.846 "data_size": 63488 00:08:28.846 } 00:08:28.846 ] 00:08:28.846 }' 00:08:28.846 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.846 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.105 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:29.105 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.105 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.105 [2024-10-29 10:57:34.515886] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:29.105 [2024-10-29 10:57:34.515931] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:29.105 [2024-10-29 10:57:34.518387] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:29.105 [2024-10-29 10:57:34.518442] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.105 [2024-10-29 10:57:34.518515] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:29.105 [2024-10-29 10:57:34.518531] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:29.105 { 00:08:29.105 "results": [ 00:08:29.105 { 00:08:29.105 "job": "raid_bdev1", 00:08:29.105 "core_mask": "0x1", 00:08:29.105 "workload": "randrw", 00:08:29.105 "percentage": 50, 00:08:29.105 "status": "finished", 00:08:29.106 "queue_depth": 1, 00:08:29.106 "io_size": 131072, 00:08:29.106 "runtime": 1.391366, 00:08:29.106 "iops": 22660.464608162052, 00:08:29.106 "mibps": 2832.5580760202565, 00:08:29.106 "io_failed": 0, 00:08:29.106 "io_timeout": 0, 00:08:29.106 "avg_latency_us": 41.64258797716, 00:08:29.106 "min_latency_us": 21.463755458515283, 00:08:29.106 "max_latency_us": 1359.3711790393013 00:08:29.106 } 00:08:29.106 ], 00:08:29.106 "core_count": 1 00:08:29.106 } 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 75043 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 75043 ']' 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 75043 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 75043 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:29.106 killing process with pid 75043 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 75043' 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 75043 00:08:29.106 [2024-10-29 10:57:34.565269] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:29.106 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 75043 00:08:29.106 [2024-10-29 10:57:34.581024] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.CmreJGJYHz 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:29.366 00:08:29.366 real 0m3.214s 00:08:29.366 user 0m4.098s 00:08:29.366 sys 0m0.526s 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:29.366 10:57:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.366 ************************************ 00:08:29.366 END TEST raid_write_error_test 00:08:29.366 ************************************ 00:08:29.366 10:57:34 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:29.366 10:57:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:29.366 10:57:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:08:29.366 10:57:34 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:29.366 10:57:34 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:29.366 10:57:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:29.366 ************************************ 00:08:29.366 START TEST raid_state_function_test 00:08:29.366 ************************************ 00:08:29.366 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 3 false 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=75176 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:29.626 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75176' 00:08:29.626 Process raid pid: 75176 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 75176 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 75176 ']' 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:29.627 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:29.627 10:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.627 [2024-10-29 10:57:34.960361] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:29.627 [2024-10-29 10:57:34.960486] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:29.627 [2024-10-29 10:57:35.110355] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:29.887 [2024-10-29 10:57:35.137011] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:29.887 [2024-10-29 10:57:35.179560] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:29.887 [2024-10-29 10:57:35.179600] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.457 [2024-10-29 10:57:35.785074] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:30.457 [2024-10-29 10:57:35.785130] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:30.457 [2024-10-29 10:57:35.785139] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:30.457 [2024-10-29 10:57:35.785149] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:30.457 [2024-10-29 10:57:35.785157] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:30.457 [2024-10-29 10:57:35.785168] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.457 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.457 "name": "Existed_Raid", 00:08:30.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.457 "strip_size_kb": 64, 00:08:30.457 "state": "configuring", 00:08:30.457 "raid_level": "raid0", 00:08:30.457 "superblock": false, 00:08:30.457 "num_base_bdevs": 3, 00:08:30.457 "num_base_bdevs_discovered": 0, 00:08:30.457 "num_base_bdevs_operational": 3, 00:08:30.457 "base_bdevs_list": [ 00:08:30.457 { 00:08:30.457 "name": "BaseBdev1", 00:08:30.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.457 "is_configured": false, 00:08:30.457 "data_offset": 0, 00:08:30.457 "data_size": 0 00:08:30.457 }, 00:08:30.457 { 00:08:30.457 "name": "BaseBdev2", 00:08:30.457 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.457 "is_configured": false, 00:08:30.457 "data_offset": 0, 00:08:30.457 "data_size": 0 00:08:30.457 }, 00:08:30.458 { 00:08:30.458 "name": "BaseBdev3", 00:08:30.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.458 "is_configured": false, 00:08:30.458 "data_offset": 0, 00:08:30.458 "data_size": 0 00:08:30.458 } 00:08:30.458 ] 00:08:30.458 }' 00:08:30.458 10:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.458 10:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.718 [2024-10-29 10:57:36.204291] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:30.718 [2024-10-29 10:57:36.204337] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.718 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.718 [2024-10-29 10:57:36.216259] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:30.718 [2024-10-29 10:57:36.216298] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:30.718 [2024-10-29 10:57:36.216307] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:30.718 [2024-10-29 10:57:36.216316] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:30.718 [2024-10-29 10:57:36.216322] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:30.718 [2024-10-29 10:57:36.216330] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.977 [2024-10-29 10:57:36.237170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:30.977 BaseBdev1 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.977 [ 00:08:30.977 { 00:08:30.977 "name": "BaseBdev1", 00:08:30.977 "aliases": [ 00:08:30.977 "3fae1be8-1e15-4720-a480-9b12490defde" 00:08:30.977 ], 00:08:30.977 "product_name": "Malloc disk", 00:08:30.977 "block_size": 512, 00:08:30.977 "num_blocks": 65536, 00:08:30.977 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:30.977 "assigned_rate_limits": { 00:08:30.977 "rw_ios_per_sec": 0, 00:08:30.977 "rw_mbytes_per_sec": 0, 00:08:30.977 "r_mbytes_per_sec": 0, 00:08:30.977 "w_mbytes_per_sec": 0 00:08:30.977 }, 00:08:30.977 "claimed": true, 00:08:30.977 "claim_type": "exclusive_write", 00:08:30.977 "zoned": false, 00:08:30.977 "supported_io_types": { 00:08:30.977 "read": true, 00:08:30.977 "write": true, 00:08:30.977 "unmap": true, 00:08:30.977 "flush": true, 00:08:30.977 "reset": true, 00:08:30.977 "nvme_admin": false, 00:08:30.977 "nvme_io": false, 00:08:30.977 "nvme_io_md": false, 00:08:30.977 "write_zeroes": true, 00:08:30.977 "zcopy": true, 00:08:30.977 "get_zone_info": false, 00:08:30.977 "zone_management": false, 00:08:30.977 "zone_append": false, 00:08:30.977 "compare": false, 00:08:30.977 "compare_and_write": false, 00:08:30.977 "abort": true, 00:08:30.977 "seek_hole": false, 00:08:30.977 "seek_data": false, 00:08:30.977 "copy": true, 00:08:30.977 "nvme_iov_md": false 00:08:30.977 }, 00:08:30.977 "memory_domains": [ 00:08:30.977 { 00:08:30.977 "dma_device_id": "system", 00:08:30.977 "dma_device_type": 1 00:08:30.977 }, 00:08:30.977 { 00:08:30.977 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:30.977 "dma_device_type": 2 00:08:30.977 } 00:08:30.977 ], 00:08:30.977 "driver_specific": {} 00:08:30.977 } 00:08:30.977 ] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.977 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.978 "name": "Existed_Raid", 00:08:30.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.978 "strip_size_kb": 64, 00:08:30.978 "state": "configuring", 00:08:30.978 "raid_level": "raid0", 00:08:30.978 "superblock": false, 00:08:30.978 "num_base_bdevs": 3, 00:08:30.978 "num_base_bdevs_discovered": 1, 00:08:30.978 "num_base_bdevs_operational": 3, 00:08:30.978 "base_bdevs_list": [ 00:08:30.978 { 00:08:30.978 "name": "BaseBdev1", 00:08:30.978 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:30.978 "is_configured": true, 00:08:30.978 "data_offset": 0, 00:08:30.978 "data_size": 65536 00:08:30.978 }, 00:08:30.978 { 00:08:30.978 "name": "BaseBdev2", 00:08:30.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.978 "is_configured": false, 00:08:30.978 "data_offset": 0, 00:08:30.978 "data_size": 0 00:08:30.978 }, 00:08:30.978 { 00:08:30.978 "name": "BaseBdev3", 00:08:30.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.978 "is_configured": false, 00:08:30.978 "data_offset": 0, 00:08:30.978 "data_size": 0 00:08:30.978 } 00:08:30.978 ] 00:08:30.978 }' 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.978 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.238 [2024-10-29 10:57:36.636524] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:31.238 [2024-10-29 10:57:36.636581] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.238 [2024-10-29 10:57:36.644529] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:31.238 [2024-10-29 10:57:36.646400] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:31.238 [2024-10-29 10:57:36.646437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:31.238 [2024-10-29 10:57:36.646446] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:31.238 [2024-10-29 10:57:36.646455] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.238 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.239 "name": "Existed_Raid", 00:08:31.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.239 "strip_size_kb": 64, 00:08:31.239 "state": "configuring", 00:08:31.239 "raid_level": "raid0", 00:08:31.239 "superblock": false, 00:08:31.239 "num_base_bdevs": 3, 00:08:31.239 "num_base_bdevs_discovered": 1, 00:08:31.239 "num_base_bdevs_operational": 3, 00:08:31.239 "base_bdevs_list": [ 00:08:31.239 { 00:08:31.239 "name": "BaseBdev1", 00:08:31.239 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:31.239 "is_configured": true, 00:08:31.239 "data_offset": 0, 00:08:31.239 "data_size": 65536 00:08:31.239 }, 00:08:31.239 { 00:08:31.239 "name": "BaseBdev2", 00:08:31.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.239 "is_configured": false, 00:08:31.239 "data_offset": 0, 00:08:31.239 "data_size": 0 00:08:31.239 }, 00:08:31.239 { 00:08:31.239 "name": "BaseBdev3", 00:08:31.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.239 "is_configured": false, 00:08:31.239 "data_offset": 0, 00:08:31.239 "data_size": 0 00:08:31.239 } 00:08:31.239 ] 00:08:31.239 }' 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.239 10:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.813 [2024-10-29 10:57:37.142938] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:31.813 BaseBdev2 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.813 [ 00:08:31.813 { 00:08:31.813 "name": "BaseBdev2", 00:08:31.813 "aliases": [ 00:08:31.813 "7be53168-fa40-47fc-8a3b-238b57a4e2d8" 00:08:31.813 ], 00:08:31.813 "product_name": "Malloc disk", 00:08:31.813 "block_size": 512, 00:08:31.813 "num_blocks": 65536, 00:08:31.813 "uuid": "7be53168-fa40-47fc-8a3b-238b57a4e2d8", 00:08:31.813 "assigned_rate_limits": { 00:08:31.813 "rw_ios_per_sec": 0, 00:08:31.813 "rw_mbytes_per_sec": 0, 00:08:31.813 "r_mbytes_per_sec": 0, 00:08:31.813 "w_mbytes_per_sec": 0 00:08:31.813 }, 00:08:31.813 "claimed": true, 00:08:31.813 "claim_type": "exclusive_write", 00:08:31.813 "zoned": false, 00:08:31.813 "supported_io_types": { 00:08:31.813 "read": true, 00:08:31.813 "write": true, 00:08:31.813 "unmap": true, 00:08:31.813 "flush": true, 00:08:31.813 "reset": true, 00:08:31.813 "nvme_admin": false, 00:08:31.813 "nvme_io": false, 00:08:31.813 "nvme_io_md": false, 00:08:31.813 "write_zeroes": true, 00:08:31.813 "zcopy": true, 00:08:31.813 "get_zone_info": false, 00:08:31.813 "zone_management": false, 00:08:31.813 "zone_append": false, 00:08:31.813 "compare": false, 00:08:31.813 "compare_and_write": false, 00:08:31.813 "abort": true, 00:08:31.813 "seek_hole": false, 00:08:31.813 "seek_data": false, 00:08:31.813 "copy": true, 00:08:31.813 "nvme_iov_md": false 00:08:31.813 }, 00:08:31.813 "memory_domains": [ 00:08:31.813 { 00:08:31.813 "dma_device_id": "system", 00:08:31.813 "dma_device_type": 1 00:08:31.813 }, 00:08:31.813 { 00:08:31.813 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.813 "dma_device_type": 2 00:08:31.813 } 00:08:31.813 ], 00:08:31.813 "driver_specific": {} 00:08:31.813 } 00:08:31.813 ] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.813 "name": "Existed_Raid", 00:08:31.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.813 "strip_size_kb": 64, 00:08:31.813 "state": "configuring", 00:08:31.813 "raid_level": "raid0", 00:08:31.813 "superblock": false, 00:08:31.813 "num_base_bdevs": 3, 00:08:31.813 "num_base_bdevs_discovered": 2, 00:08:31.813 "num_base_bdevs_operational": 3, 00:08:31.813 "base_bdevs_list": [ 00:08:31.813 { 00:08:31.813 "name": "BaseBdev1", 00:08:31.813 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:31.813 "is_configured": true, 00:08:31.813 "data_offset": 0, 00:08:31.813 "data_size": 65536 00:08:31.813 }, 00:08:31.813 { 00:08:31.813 "name": "BaseBdev2", 00:08:31.813 "uuid": "7be53168-fa40-47fc-8a3b-238b57a4e2d8", 00:08:31.813 "is_configured": true, 00:08:31.813 "data_offset": 0, 00:08:31.813 "data_size": 65536 00:08:31.813 }, 00:08:31.813 { 00:08:31.813 "name": "BaseBdev3", 00:08:31.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.813 "is_configured": false, 00:08:31.813 "data_offset": 0, 00:08:31.813 "data_size": 0 00:08:31.813 } 00:08:31.813 ] 00:08:31.813 }' 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.813 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.389 [2024-10-29 10:57:37.633924] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:32.389 [2024-10-29 10:57:37.633976] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:32.389 [2024-10-29 10:57:37.633989] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:32.389 [2024-10-29 10:57:37.634314] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:32.389 [2024-10-29 10:57:37.634555] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:32.389 [2024-10-29 10:57:37.634577] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:32.389 [2024-10-29 10:57:37.634837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:32.389 BaseBdev3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.389 [ 00:08:32.389 { 00:08:32.389 "name": "BaseBdev3", 00:08:32.389 "aliases": [ 00:08:32.389 "55a6d5a5-99ae-4b29-b0c2-315b9ab788d1" 00:08:32.389 ], 00:08:32.389 "product_name": "Malloc disk", 00:08:32.389 "block_size": 512, 00:08:32.389 "num_blocks": 65536, 00:08:32.389 "uuid": "55a6d5a5-99ae-4b29-b0c2-315b9ab788d1", 00:08:32.389 "assigned_rate_limits": { 00:08:32.389 "rw_ios_per_sec": 0, 00:08:32.389 "rw_mbytes_per_sec": 0, 00:08:32.389 "r_mbytes_per_sec": 0, 00:08:32.389 "w_mbytes_per_sec": 0 00:08:32.389 }, 00:08:32.389 "claimed": true, 00:08:32.389 "claim_type": "exclusive_write", 00:08:32.389 "zoned": false, 00:08:32.389 "supported_io_types": { 00:08:32.389 "read": true, 00:08:32.389 "write": true, 00:08:32.389 "unmap": true, 00:08:32.389 "flush": true, 00:08:32.389 "reset": true, 00:08:32.389 "nvme_admin": false, 00:08:32.389 "nvme_io": false, 00:08:32.389 "nvme_io_md": false, 00:08:32.389 "write_zeroes": true, 00:08:32.389 "zcopy": true, 00:08:32.389 "get_zone_info": false, 00:08:32.389 "zone_management": false, 00:08:32.389 "zone_append": false, 00:08:32.389 "compare": false, 00:08:32.389 "compare_and_write": false, 00:08:32.389 "abort": true, 00:08:32.389 "seek_hole": false, 00:08:32.389 "seek_data": false, 00:08:32.389 "copy": true, 00:08:32.389 "nvme_iov_md": false 00:08:32.389 }, 00:08:32.389 "memory_domains": [ 00:08:32.389 { 00:08:32.389 "dma_device_id": "system", 00:08:32.389 "dma_device_type": 1 00:08:32.389 }, 00:08:32.389 { 00:08:32.389 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.389 "dma_device_type": 2 00:08:32.389 } 00:08:32.389 ], 00:08:32.389 "driver_specific": {} 00:08:32.389 } 00:08:32.389 ] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.389 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.389 "name": "Existed_Raid", 00:08:32.390 "uuid": "ca4938bc-36b1-410f-8252-6d97c61554b3", 00:08:32.390 "strip_size_kb": 64, 00:08:32.390 "state": "online", 00:08:32.390 "raid_level": "raid0", 00:08:32.390 "superblock": false, 00:08:32.390 "num_base_bdevs": 3, 00:08:32.390 "num_base_bdevs_discovered": 3, 00:08:32.390 "num_base_bdevs_operational": 3, 00:08:32.390 "base_bdevs_list": [ 00:08:32.390 { 00:08:32.390 "name": "BaseBdev1", 00:08:32.390 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:32.390 "is_configured": true, 00:08:32.390 "data_offset": 0, 00:08:32.390 "data_size": 65536 00:08:32.390 }, 00:08:32.390 { 00:08:32.390 "name": "BaseBdev2", 00:08:32.390 "uuid": "7be53168-fa40-47fc-8a3b-238b57a4e2d8", 00:08:32.390 "is_configured": true, 00:08:32.390 "data_offset": 0, 00:08:32.390 "data_size": 65536 00:08:32.390 }, 00:08:32.390 { 00:08:32.390 "name": "BaseBdev3", 00:08:32.390 "uuid": "55a6d5a5-99ae-4b29-b0c2-315b9ab788d1", 00:08:32.390 "is_configured": true, 00:08:32.390 "data_offset": 0, 00:08:32.390 "data_size": 65536 00:08:32.390 } 00:08:32.390 ] 00:08:32.390 }' 00:08:32.390 10:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.390 10:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.650 [2024-10-29 10:57:38.089525] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:32.650 "name": "Existed_Raid", 00:08:32.650 "aliases": [ 00:08:32.650 "ca4938bc-36b1-410f-8252-6d97c61554b3" 00:08:32.650 ], 00:08:32.650 "product_name": "Raid Volume", 00:08:32.650 "block_size": 512, 00:08:32.650 "num_blocks": 196608, 00:08:32.650 "uuid": "ca4938bc-36b1-410f-8252-6d97c61554b3", 00:08:32.650 "assigned_rate_limits": { 00:08:32.650 "rw_ios_per_sec": 0, 00:08:32.650 "rw_mbytes_per_sec": 0, 00:08:32.650 "r_mbytes_per_sec": 0, 00:08:32.650 "w_mbytes_per_sec": 0 00:08:32.650 }, 00:08:32.650 "claimed": false, 00:08:32.650 "zoned": false, 00:08:32.650 "supported_io_types": { 00:08:32.650 "read": true, 00:08:32.650 "write": true, 00:08:32.650 "unmap": true, 00:08:32.650 "flush": true, 00:08:32.650 "reset": true, 00:08:32.650 "nvme_admin": false, 00:08:32.650 "nvme_io": false, 00:08:32.650 "nvme_io_md": false, 00:08:32.650 "write_zeroes": true, 00:08:32.650 "zcopy": false, 00:08:32.650 "get_zone_info": false, 00:08:32.650 "zone_management": false, 00:08:32.650 "zone_append": false, 00:08:32.650 "compare": false, 00:08:32.650 "compare_and_write": false, 00:08:32.650 "abort": false, 00:08:32.650 "seek_hole": false, 00:08:32.650 "seek_data": false, 00:08:32.650 "copy": false, 00:08:32.650 "nvme_iov_md": false 00:08:32.650 }, 00:08:32.650 "memory_domains": [ 00:08:32.650 { 00:08:32.650 "dma_device_id": "system", 00:08:32.650 "dma_device_type": 1 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.650 "dma_device_type": 2 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "dma_device_id": "system", 00:08:32.650 "dma_device_type": 1 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.650 "dma_device_type": 2 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "dma_device_id": "system", 00:08:32.650 "dma_device_type": 1 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.650 "dma_device_type": 2 00:08:32.650 } 00:08:32.650 ], 00:08:32.650 "driver_specific": { 00:08:32.650 "raid": { 00:08:32.650 "uuid": "ca4938bc-36b1-410f-8252-6d97c61554b3", 00:08:32.650 "strip_size_kb": 64, 00:08:32.650 "state": "online", 00:08:32.650 "raid_level": "raid0", 00:08:32.650 "superblock": false, 00:08:32.650 "num_base_bdevs": 3, 00:08:32.650 "num_base_bdevs_discovered": 3, 00:08:32.650 "num_base_bdevs_operational": 3, 00:08:32.650 "base_bdevs_list": [ 00:08:32.650 { 00:08:32.650 "name": "BaseBdev1", 00:08:32.650 "uuid": "3fae1be8-1e15-4720-a480-9b12490defde", 00:08:32.650 "is_configured": true, 00:08:32.650 "data_offset": 0, 00:08:32.650 "data_size": 65536 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "name": "BaseBdev2", 00:08:32.650 "uuid": "7be53168-fa40-47fc-8a3b-238b57a4e2d8", 00:08:32.650 "is_configured": true, 00:08:32.650 "data_offset": 0, 00:08:32.650 "data_size": 65536 00:08:32.650 }, 00:08:32.650 { 00:08:32.650 "name": "BaseBdev3", 00:08:32.650 "uuid": "55a6d5a5-99ae-4b29-b0c2-315b9ab788d1", 00:08:32.650 "is_configured": true, 00:08:32.650 "data_offset": 0, 00:08:32.650 "data_size": 65536 00:08:32.650 } 00:08:32.650 ] 00:08:32.650 } 00:08:32.650 } 00:08:32.650 }' 00:08:32.650 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:32.911 BaseBdev2 00:08:32.911 BaseBdev3' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:32.911 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.912 [2024-10-29 10:57:38.352753] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:32.912 [2024-10-29 10:57:38.352785] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:32.912 [2024-10-29 10:57:38.352845] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.912 "name": "Existed_Raid", 00:08:32.912 "uuid": "ca4938bc-36b1-410f-8252-6d97c61554b3", 00:08:32.912 "strip_size_kb": 64, 00:08:32.912 "state": "offline", 00:08:32.912 "raid_level": "raid0", 00:08:32.912 "superblock": false, 00:08:32.912 "num_base_bdevs": 3, 00:08:32.912 "num_base_bdevs_discovered": 2, 00:08:32.912 "num_base_bdevs_operational": 2, 00:08:32.912 "base_bdevs_list": [ 00:08:32.912 { 00:08:32.912 "name": null, 00:08:32.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.912 "is_configured": false, 00:08:32.912 "data_offset": 0, 00:08:32.912 "data_size": 65536 00:08:32.912 }, 00:08:32.912 { 00:08:32.912 "name": "BaseBdev2", 00:08:32.912 "uuid": "7be53168-fa40-47fc-8a3b-238b57a4e2d8", 00:08:32.912 "is_configured": true, 00:08:32.912 "data_offset": 0, 00:08:32.912 "data_size": 65536 00:08:32.912 }, 00:08:32.912 { 00:08:32.912 "name": "BaseBdev3", 00:08:32.912 "uuid": "55a6d5a5-99ae-4b29-b0c2-315b9ab788d1", 00:08:32.912 "is_configured": true, 00:08:32.912 "data_offset": 0, 00:08:32.912 "data_size": 65536 00:08:32.912 } 00:08:32.912 ] 00:08:32.912 }' 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.912 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 [2024-10-29 10:57:38.779267] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 [2024-10-29 10:57:38.834416] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:33.483 [2024-10-29 10:57:38.834476] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 [ 00:08:33.483 { 00:08:33.483 "name": "BaseBdev2", 00:08:33.483 "aliases": [ 00:08:33.483 "b1a628a7-0d9e-4ea2-a149-523dd9674e99" 00:08:33.483 ], 00:08:33.483 "product_name": "Malloc disk", 00:08:33.483 "block_size": 512, 00:08:33.483 "num_blocks": 65536, 00:08:33.483 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:33.483 "assigned_rate_limits": { 00:08:33.483 "rw_ios_per_sec": 0, 00:08:33.483 "rw_mbytes_per_sec": 0, 00:08:33.483 "r_mbytes_per_sec": 0, 00:08:33.483 "w_mbytes_per_sec": 0 00:08:33.483 }, 00:08:33.483 "claimed": false, 00:08:33.483 "zoned": false, 00:08:33.483 "supported_io_types": { 00:08:33.483 "read": true, 00:08:33.483 "write": true, 00:08:33.483 "unmap": true, 00:08:33.483 "flush": true, 00:08:33.483 "reset": true, 00:08:33.483 "nvme_admin": false, 00:08:33.483 "nvme_io": false, 00:08:33.483 "nvme_io_md": false, 00:08:33.483 "write_zeroes": true, 00:08:33.483 "zcopy": true, 00:08:33.483 "get_zone_info": false, 00:08:33.483 "zone_management": false, 00:08:33.483 "zone_append": false, 00:08:33.483 "compare": false, 00:08:33.483 "compare_and_write": false, 00:08:33.483 "abort": true, 00:08:33.483 "seek_hole": false, 00:08:33.483 "seek_data": false, 00:08:33.483 "copy": true, 00:08:33.483 "nvme_iov_md": false 00:08:33.483 }, 00:08:33.483 "memory_domains": [ 00:08:33.483 { 00:08:33.483 "dma_device_id": "system", 00:08:33.483 "dma_device_type": 1 00:08:33.483 }, 00:08:33.483 { 00:08:33.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.483 "dma_device_type": 2 00:08:33.483 } 00:08:33.483 ], 00:08:33.483 "driver_specific": {} 00:08:33.483 } 00:08:33.483 ] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 BaseBdev3 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.483 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.484 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:33.484 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.484 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.744 [ 00:08:33.744 { 00:08:33.744 "name": "BaseBdev3", 00:08:33.744 "aliases": [ 00:08:33.744 "38d6b9e9-cbd7-44bb-918e-beb5d9474429" 00:08:33.744 ], 00:08:33.744 "product_name": "Malloc disk", 00:08:33.744 "block_size": 512, 00:08:33.744 "num_blocks": 65536, 00:08:33.744 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:33.744 "assigned_rate_limits": { 00:08:33.744 "rw_ios_per_sec": 0, 00:08:33.744 "rw_mbytes_per_sec": 0, 00:08:33.744 "r_mbytes_per_sec": 0, 00:08:33.744 "w_mbytes_per_sec": 0 00:08:33.744 }, 00:08:33.744 "claimed": false, 00:08:33.744 "zoned": false, 00:08:33.744 "supported_io_types": { 00:08:33.744 "read": true, 00:08:33.744 "write": true, 00:08:33.744 "unmap": true, 00:08:33.744 "flush": true, 00:08:33.744 "reset": true, 00:08:33.744 "nvme_admin": false, 00:08:33.744 "nvme_io": false, 00:08:33.744 "nvme_io_md": false, 00:08:33.744 "write_zeroes": true, 00:08:33.744 "zcopy": true, 00:08:33.744 "get_zone_info": false, 00:08:33.744 "zone_management": false, 00:08:33.744 "zone_append": false, 00:08:33.744 "compare": false, 00:08:33.744 "compare_and_write": false, 00:08:33.744 "abort": true, 00:08:33.744 "seek_hole": false, 00:08:33.744 "seek_data": false, 00:08:33.744 "copy": true, 00:08:33.744 "nvme_iov_md": false 00:08:33.744 }, 00:08:33.744 "memory_domains": [ 00:08:33.744 { 00:08:33.744 "dma_device_id": "system", 00:08:33.744 "dma_device_type": 1 00:08:33.744 }, 00:08:33.744 { 00:08:33.744 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.744 "dma_device_type": 2 00:08:33.744 } 00:08:33.744 ], 00:08:33.744 "driver_specific": {} 00:08:33.744 } 00:08:33.744 ] 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.744 10:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.744 [2024-10-29 10:57:38.999461] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:33.744 [2024-10-29 10:57:38.999516] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:33.744 [2024-10-29 10:57:38.999545] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:33.744 [2024-10-29 10:57:39.001455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.744 "name": "Existed_Raid", 00:08:33.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:33.744 "strip_size_kb": 64, 00:08:33.744 "state": "configuring", 00:08:33.744 "raid_level": "raid0", 00:08:33.744 "superblock": false, 00:08:33.744 "num_base_bdevs": 3, 00:08:33.744 "num_base_bdevs_discovered": 2, 00:08:33.744 "num_base_bdevs_operational": 3, 00:08:33.744 "base_bdevs_list": [ 00:08:33.744 { 00:08:33.744 "name": "BaseBdev1", 00:08:33.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:33.744 "is_configured": false, 00:08:33.744 "data_offset": 0, 00:08:33.744 "data_size": 0 00:08:33.744 }, 00:08:33.744 { 00:08:33.744 "name": "BaseBdev2", 00:08:33.744 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:33.744 "is_configured": true, 00:08:33.744 "data_offset": 0, 00:08:33.744 "data_size": 65536 00:08:33.744 }, 00:08:33.744 { 00:08:33.744 "name": "BaseBdev3", 00:08:33.744 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:33.744 "is_configured": true, 00:08:33.744 "data_offset": 0, 00:08:33.744 "data_size": 65536 00:08:33.744 } 00:08:33.744 ] 00:08:33.744 }' 00:08:33.744 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.745 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.004 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:34.004 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.004 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.004 [2024-10-29 10:57:39.502551] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.264 "name": "Existed_Raid", 00:08:34.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.264 "strip_size_kb": 64, 00:08:34.264 "state": "configuring", 00:08:34.264 "raid_level": "raid0", 00:08:34.264 "superblock": false, 00:08:34.264 "num_base_bdevs": 3, 00:08:34.264 "num_base_bdevs_discovered": 1, 00:08:34.264 "num_base_bdevs_operational": 3, 00:08:34.264 "base_bdevs_list": [ 00:08:34.264 { 00:08:34.264 "name": "BaseBdev1", 00:08:34.264 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.264 "is_configured": false, 00:08:34.264 "data_offset": 0, 00:08:34.264 "data_size": 0 00:08:34.264 }, 00:08:34.264 { 00:08:34.264 "name": null, 00:08:34.264 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:34.264 "is_configured": false, 00:08:34.264 "data_offset": 0, 00:08:34.264 "data_size": 65536 00:08:34.264 }, 00:08:34.264 { 00:08:34.264 "name": "BaseBdev3", 00:08:34.264 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:34.264 "is_configured": true, 00:08:34.264 "data_offset": 0, 00:08:34.264 "data_size": 65536 00:08:34.264 } 00:08:34.264 ] 00:08:34.264 }' 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.264 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:34.523 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 [2024-10-29 10:57:39.976718] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:34.524 BaseBdev1 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 10:57:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.524 [ 00:08:34.524 { 00:08:34.524 "name": "BaseBdev1", 00:08:34.524 "aliases": [ 00:08:34.524 "74573f7f-09be-4159-a646-2cf1386d1cd4" 00:08:34.524 ], 00:08:34.524 "product_name": "Malloc disk", 00:08:34.524 "block_size": 512, 00:08:34.524 "num_blocks": 65536, 00:08:34.524 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:34.524 "assigned_rate_limits": { 00:08:34.524 "rw_ios_per_sec": 0, 00:08:34.524 "rw_mbytes_per_sec": 0, 00:08:34.524 "r_mbytes_per_sec": 0, 00:08:34.524 "w_mbytes_per_sec": 0 00:08:34.524 }, 00:08:34.524 "claimed": true, 00:08:34.524 "claim_type": "exclusive_write", 00:08:34.524 "zoned": false, 00:08:34.524 "supported_io_types": { 00:08:34.524 "read": true, 00:08:34.524 "write": true, 00:08:34.524 "unmap": true, 00:08:34.524 "flush": true, 00:08:34.524 "reset": true, 00:08:34.524 "nvme_admin": false, 00:08:34.524 "nvme_io": false, 00:08:34.524 "nvme_io_md": false, 00:08:34.524 "write_zeroes": true, 00:08:34.524 "zcopy": true, 00:08:34.524 "get_zone_info": false, 00:08:34.524 "zone_management": false, 00:08:34.524 "zone_append": false, 00:08:34.524 "compare": false, 00:08:34.524 "compare_and_write": false, 00:08:34.524 "abort": true, 00:08:34.524 "seek_hole": false, 00:08:34.524 "seek_data": false, 00:08:34.524 "copy": true, 00:08:34.524 "nvme_iov_md": false 00:08:34.524 }, 00:08:34.524 "memory_domains": [ 00:08:34.524 { 00:08:34.524 "dma_device_id": "system", 00:08:34.524 "dma_device_type": 1 00:08:34.524 }, 00:08:34.524 { 00:08:34.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.524 "dma_device_type": 2 00:08:34.524 } 00:08:34.524 ], 00:08:34.524 "driver_specific": {} 00:08:34.524 } 00:08:34.524 ] 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.524 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.783 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.783 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.783 "name": "Existed_Raid", 00:08:34.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.783 "strip_size_kb": 64, 00:08:34.783 "state": "configuring", 00:08:34.783 "raid_level": "raid0", 00:08:34.783 "superblock": false, 00:08:34.783 "num_base_bdevs": 3, 00:08:34.783 "num_base_bdevs_discovered": 2, 00:08:34.783 "num_base_bdevs_operational": 3, 00:08:34.783 "base_bdevs_list": [ 00:08:34.783 { 00:08:34.783 "name": "BaseBdev1", 00:08:34.783 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:34.783 "is_configured": true, 00:08:34.783 "data_offset": 0, 00:08:34.783 "data_size": 65536 00:08:34.783 }, 00:08:34.783 { 00:08:34.783 "name": null, 00:08:34.783 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:34.783 "is_configured": false, 00:08:34.783 "data_offset": 0, 00:08:34.783 "data_size": 65536 00:08:34.783 }, 00:08:34.783 { 00:08:34.783 "name": "BaseBdev3", 00:08:34.783 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:34.783 "is_configured": true, 00:08:34.783 "data_offset": 0, 00:08:34.783 "data_size": 65536 00:08:34.783 } 00:08:34.783 ] 00:08:34.783 }' 00:08:34.784 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.784 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.043 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.043 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.044 [2024-10-29 10:57:40.519851] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.044 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.303 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.303 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.303 "name": "Existed_Raid", 00:08:35.303 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.303 "strip_size_kb": 64, 00:08:35.303 "state": "configuring", 00:08:35.303 "raid_level": "raid0", 00:08:35.303 "superblock": false, 00:08:35.303 "num_base_bdevs": 3, 00:08:35.303 "num_base_bdevs_discovered": 1, 00:08:35.303 "num_base_bdevs_operational": 3, 00:08:35.303 "base_bdevs_list": [ 00:08:35.303 { 00:08:35.303 "name": "BaseBdev1", 00:08:35.303 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:35.303 "is_configured": true, 00:08:35.303 "data_offset": 0, 00:08:35.303 "data_size": 65536 00:08:35.303 }, 00:08:35.303 { 00:08:35.303 "name": null, 00:08:35.303 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:35.303 "is_configured": false, 00:08:35.303 "data_offset": 0, 00:08:35.303 "data_size": 65536 00:08:35.303 }, 00:08:35.303 { 00:08:35.303 "name": null, 00:08:35.303 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:35.303 "is_configured": false, 00:08:35.303 "data_offset": 0, 00:08:35.303 "data_size": 65536 00:08:35.303 } 00:08:35.303 ] 00:08:35.303 }' 00:08:35.303 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.303 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.563 [2024-10-29 10:57:40.951363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.563 "name": "Existed_Raid", 00:08:35.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.563 "strip_size_kb": 64, 00:08:35.563 "state": "configuring", 00:08:35.563 "raid_level": "raid0", 00:08:35.563 "superblock": false, 00:08:35.563 "num_base_bdevs": 3, 00:08:35.563 "num_base_bdevs_discovered": 2, 00:08:35.563 "num_base_bdevs_operational": 3, 00:08:35.563 "base_bdevs_list": [ 00:08:35.563 { 00:08:35.563 "name": "BaseBdev1", 00:08:35.563 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:35.563 "is_configured": true, 00:08:35.563 "data_offset": 0, 00:08:35.563 "data_size": 65536 00:08:35.563 }, 00:08:35.563 { 00:08:35.563 "name": null, 00:08:35.563 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:35.563 "is_configured": false, 00:08:35.563 "data_offset": 0, 00:08:35.563 "data_size": 65536 00:08:35.563 }, 00:08:35.563 { 00:08:35.563 "name": "BaseBdev3", 00:08:35.563 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:35.563 "is_configured": true, 00:08:35.563 "data_offset": 0, 00:08:35.563 "data_size": 65536 00:08:35.563 } 00:08:35.563 ] 00:08:35.563 }' 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.563 10:57:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.132 [2024-10-29 10:57:41.426751] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.132 "name": "Existed_Raid", 00:08:36.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.132 "strip_size_kb": 64, 00:08:36.132 "state": "configuring", 00:08:36.132 "raid_level": "raid0", 00:08:36.132 "superblock": false, 00:08:36.132 "num_base_bdevs": 3, 00:08:36.132 "num_base_bdevs_discovered": 1, 00:08:36.132 "num_base_bdevs_operational": 3, 00:08:36.132 "base_bdevs_list": [ 00:08:36.132 { 00:08:36.132 "name": null, 00:08:36.132 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:36.132 "is_configured": false, 00:08:36.132 "data_offset": 0, 00:08:36.132 "data_size": 65536 00:08:36.132 }, 00:08:36.132 { 00:08:36.132 "name": null, 00:08:36.132 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:36.132 "is_configured": false, 00:08:36.132 "data_offset": 0, 00:08:36.132 "data_size": 65536 00:08:36.132 }, 00:08:36.132 { 00:08:36.132 "name": "BaseBdev3", 00:08:36.132 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:36.132 "is_configured": true, 00:08:36.132 "data_offset": 0, 00:08:36.132 "data_size": 65536 00:08:36.132 } 00:08:36.132 ] 00:08:36.132 }' 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.132 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.391 [2024-10-29 10:57:41.862184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.391 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.650 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.650 "name": "Existed_Raid", 00:08:36.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.650 "strip_size_kb": 64, 00:08:36.650 "state": "configuring", 00:08:36.650 "raid_level": "raid0", 00:08:36.650 "superblock": false, 00:08:36.650 "num_base_bdevs": 3, 00:08:36.650 "num_base_bdevs_discovered": 2, 00:08:36.650 "num_base_bdevs_operational": 3, 00:08:36.650 "base_bdevs_list": [ 00:08:36.650 { 00:08:36.650 "name": null, 00:08:36.650 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:36.650 "is_configured": false, 00:08:36.650 "data_offset": 0, 00:08:36.650 "data_size": 65536 00:08:36.650 }, 00:08:36.650 { 00:08:36.650 "name": "BaseBdev2", 00:08:36.650 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:36.650 "is_configured": true, 00:08:36.650 "data_offset": 0, 00:08:36.650 "data_size": 65536 00:08:36.650 }, 00:08:36.650 { 00:08:36.650 "name": "BaseBdev3", 00:08:36.650 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:36.650 "is_configured": true, 00:08:36.650 "data_offset": 0, 00:08:36.650 "data_size": 65536 00:08:36.650 } 00:08:36.650 ] 00:08:36.650 }' 00:08:36.650 10:57:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.650 10:57:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 74573f7f-09be-4159-a646-2cf1386d1cd4 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.908 [2024-10-29 10:57:42.390531] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:36.908 [2024-10-29 10:57:42.390580] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:36.908 [2024-10-29 10:57:42.390592] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:36.908 [2024-10-29 10:57:42.390926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:36.908 [2024-10-29 10:57:42.391088] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:36.908 [2024-10-29 10:57:42.391102] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:36.908 [2024-10-29 10:57:42.391334] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:36.908 NewBaseBdev 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:08:36.908 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.909 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.168 [ 00:08:37.168 { 00:08:37.168 "name": "NewBaseBdev", 00:08:37.168 "aliases": [ 00:08:37.168 "74573f7f-09be-4159-a646-2cf1386d1cd4" 00:08:37.168 ], 00:08:37.168 "product_name": "Malloc disk", 00:08:37.168 "block_size": 512, 00:08:37.168 "num_blocks": 65536, 00:08:37.168 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:37.168 "assigned_rate_limits": { 00:08:37.168 "rw_ios_per_sec": 0, 00:08:37.168 "rw_mbytes_per_sec": 0, 00:08:37.168 "r_mbytes_per_sec": 0, 00:08:37.168 "w_mbytes_per_sec": 0 00:08:37.168 }, 00:08:37.168 "claimed": true, 00:08:37.168 "claim_type": "exclusive_write", 00:08:37.168 "zoned": false, 00:08:37.168 "supported_io_types": { 00:08:37.168 "read": true, 00:08:37.168 "write": true, 00:08:37.168 "unmap": true, 00:08:37.168 "flush": true, 00:08:37.168 "reset": true, 00:08:37.168 "nvme_admin": false, 00:08:37.168 "nvme_io": false, 00:08:37.168 "nvme_io_md": false, 00:08:37.168 "write_zeroes": true, 00:08:37.168 "zcopy": true, 00:08:37.168 "get_zone_info": false, 00:08:37.168 "zone_management": false, 00:08:37.168 "zone_append": false, 00:08:37.168 "compare": false, 00:08:37.168 "compare_and_write": false, 00:08:37.168 "abort": true, 00:08:37.168 "seek_hole": false, 00:08:37.168 "seek_data": false, 00:08:37.168 "copy": true, 00:08:37.168 "nvme_iov_md": false 00:08:37.168 }, 00:08:37.168 "memory_domains": [ 00:08:37.168 { 00:08:37.168 "dma_device_id": "system", 00:08:37.168 "dma_device_type": 1 00:08:37.168 }, 00:08:37.168 { 00:08:37.168 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:37.168 "dma_device_type": 2 00:08:37.168 } 00:08:37.168 ], 00:08:37.168 "driver_specific": {} 00:08:37.168 } 00:08:37.168 ] 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.168 "name": "Existed_Raid", 00:08:37.168 "uuid": "6bc1a091-a794-403b-9282-4bb3887b4b88", 00:08:37.168 "strip_size_kb": 64, 00:08:37.168 "state": "online", 00:08:37.168 "raid_level": "raid0", 00:08:37.168 "superblock": false, 00:08:37.168 "num_base_bdevs": 3, 00:08:37.168 "num_base_bdevs_discovered": 3, 00:08:37.168 "num_base_bdevs_operational": 3, 00:08:37.168 "base_bdevs_list": [ 00:08:37.168 { 00:08:37.168 "name": "NewBaseBdev", 00:08:37.168 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:37.168 "is_configured": true, 00:08:37.168 "data_offset": 0, 00:08:37.168 "data_size": 65536 00:08:37.168 }, 00:08:37.168 { 00:08:37.168 "name": "BaseBdev2", 00:08:37.168 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:37.168 "is_configured": true, 00:08:37.168 "data_offset": 0, 00:08:37.168 "data_size": 65536 00:08:37.168 }, 00:08:37.168 { 00:08:37.168 "name": "BaseBdev3", 00:08:37.168 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:37.168 "is_configured": true, 00:08:37.168 "data_offset": 0, 00:08:37.168 "data_size": 65536 00:08:37.168 } 00:08:37.168 ] 00:08:37.168 }' 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.168 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.426 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.426 [2024-10-29 10:57:42.910096] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:37.684 10:57:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.684 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:37.684 "name": "Existed_Raid", 00:08:37.684 "aliases": [ 00:08:37.684 "6bc1a091-a794-403b-9282-4bb3887b4b88" 00:08:37.684 ], 00:08:37.684 "product_name": "Raid Volume", 00:08:37.684 "block_size": 512, 00:08:37.684 "num_blocks": 196608, 00:08:37.684 "uuid": "6bc1a091-a794-403b-9282-4bb3887b4b88", 00:08:37.684 "assigned_rate_limits": { 00:08:37.684 "rw_ios_per_sec": 0, 00:08:37.684 "rw_mbytes_per_sec": 0, 00:08:37.684 "r_mbytes_per_sec": 0, 00:08:37.684 "w_mbytes_per_sec": 0 00:08:37.684 }, 00:08:37.684 "claimed": false, 00:08:37.684 "zoned": false, 00:08:37.684 "supported_io_types": { 00:08:37.684 "read": true, 00:08:37.684 "write": true, 00:08:37.684 "unmap": true, 00:08:37.684 "flush": true, 00:08:37.684 "reset": true, 00:08:37.684 "nvme_admin": false, 00:08:37.684 "nvme_io": false, 00:08:37.684 "nvme_io_md": false, 00:08:37.684 "write_zeroes": true, 00:08:37.684 "zcopy": false, 00:08:37.684 "get_zone_info": false, 00:08:37.684 "zone_management": false, 00:08:37.684 "zone_append": false, 00:08:37.684 "compare": false, 00:08:37.684 "compare_and_write": false, 00:08:37.684 "abort": false, 00:08:37.684 "seek_hole": false, 00:08:37.684 "seek_data": false, 00:08:37.684 "copy": false, 00:08:37.684 "nvme_iov_md": false 00:08:37.684 }, 00:08:37.684 "memory_domains": [ 00:08:37.684 { 00:08:37.684 "dma_device_id": "system", 00:08:37.684 "dma_device_type": 1 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:37.685 "dma_device_type": 2 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "dma_device_id": "system", 00:08:37.685 "dma_device_type": 1 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:37.685 "dma_device_type": 2 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "dma_device_id": "system", 00:08:37.685 "dma_device_type": 1 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:37.685 "dma_device_type": 2 00:08:37.685 } 00:08:37.685 ], 00:08:37.685 "driver_specific": { 00:08:37.685 "raid": { 00:08:37.685 "uuid": "6bc1a091-a794-403b-9282-4bb3887b4b88", 00:08:37.685 "strip_size_kb": 64, 00:08:37.685 "state": "online", 00:08:37.685 "raid_level": "raid0", 00:08:37.685 "superblock": false, 00:08:37.685 "num_base_bdevs": 3, 00:08:37.685 "num_base_bdevs_discovered": 3, 00:08:37.685 "num_base_bdevs_operational": 3, 00:08:37.685 "base_bdevs_list": [ 00:08:37.685 { 00:08:37.685 "name": "NewBaseBdev", 00:08:37.685 "uuid": "74573f7f-09be-4159-a646-2cf1386d1cd4", 00:08:37.685 "is_configured": true, 00:08:37.685 "data_offset": 0, 00:08:37.685 "data_size": 65536 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "name": "BaseBdev2", 00:08:37.685 "uuid": "b1a628a7-0d9e-4ea2-a149-523dd9674e99", 00:08:37.685 "is_configured": true, 00:08:37.685 "data_offset": 0, 00:08:37.685 "data_size": 65536 00:08:37.685 }, 00:08:37.685 { 00:08:37.685 "name": "BaseBdev3", 00:08:37.685 "uuid": "38d6b9e9-cbd7-44bb-918e-beb5d9474429", 00:08:37.685 "is_configured": true, 00:08:37.685 "data_offset": 0, 00:08:37.685 "data_size": 65536 00:08:37.685 } 00:08:37.685 ] 00:08:37.685 } 00:08:37.685 } 00:08:37.685 }' 00:08:37.685 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:37.685 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:37.685 BaseBdev2 00:08:37.685 BaseBdev3' 00:08:37.685 10:57:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:37.685 [2024-10-29 10:57:43.157295] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:37.685 [2024-10-29 10:57:43.157342] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:37.685 [2024-10-29 10:57:43.157460] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:37.685 [2024-10-29 10:57:43.157526] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:37.685 [2024-10-29 10:57:43.157545] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 75176 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 75176 ']' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 75176 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:37.685 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 75176 00:08:37.944 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:37.944 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:37.944 killing process with pid 75176 00:08:37.944 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 75176' 00:08:37.944 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 75176 00:08:37.944 [2024-10-29 10:57:43.206569] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:37.944 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 75176 00:08:37.944 [2024-10-29 10:57:43.265965] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:38.203 00:08:38.203 real 0m8.728s 00:08:38.203 user 0m14.850s 00:08:38.203 sys 0m1.663s 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.203 ************************************ 00:08:38.203 END TEST raid_state_function_test 00:08:38.203 ************************************ 00:08:38.203 10:57:43 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:38.203 10:57:43 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:38.203 10:57:43 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:38.203 10:57:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:38.203 ************************************ 00:08:38.203 START TEST raid_state_function_test_sb 00:08:38.203 ************************************ 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 3 true 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:38.203 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75780 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:38.204 Process raid pid: 75780 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75780' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75780 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 75780 ']' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:38.204 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:38.204 10:57:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.463 [2024-10-29 10:57:43.757445] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:38.463 [2024-10-29 10:57:43.757582] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:38.463 [2024-10-29 10:57:43.925294] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:38.721 [2024-10-29 10:57:43.970744] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:38.721 [2024-10-29 10:57:44.048932] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:38.721 [2024-10-29 10:57:44.048974] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.291 [2024-10-29 10:57:44.586456] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.291 [2024-10-29 10:57:44.586518] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.291 [2024-10-29 10:57:44.586537] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.291 [2024-10-29 10:57:44.586549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.291 [2024-10-29 10:57:44.586558] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:39.291 [2024-10-29 10:57:44.586570] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.291 "name": "Existed_Raid", 00:08:39.291 "uuid": "47b4201a-8af8-47c8-9f41-8e87cf5a44a6", 00:08:39.291 "strip_size_kb": 64, 00:08:39.291 "state": "configuring", 00:08:39.291 "raid_level": "raid0", 00:08:39.291 "superblock": true, 00:08:39.291 "num_base_bdevs": 3, 00:08:39.291 "num_base_bdevs_discovered": 0, 00:08:39.291 "num_base_bdevs_operational": 3, 00:08:39.291 "base_bdevs_list": [ 00:08:39.291 { 00:08:39.291 "name": "BaseBdev1", 00:08:39.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.291 "is_configured": false, 00:08:39.291 "data_offset": 0, 00:08:39.291 "data_size": 0 00:08:39.291 }, 00:08:39.291 { 00:08:39.291 "name": "BaseBdev2", 00:08:39.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.291 "is_configured": false, 00:08:39.291 "data_offset": 0, 00:08:39.291 "data_size": 0 00:08:39.291 }, 00:08:39.291 { 00:08:39.291 "name": "BaseBdev3", 00:08:39.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.291 "is_configured": false, 00:08:39.291 "data_offset": 0, 00:08:39.291 "data_size": 0 00:08:39.291 } 00:08:39.291 ] 00:08:39.291 }' 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.291 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 [2024-10-29 10:57:44.953754] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:39.552 [2024-10-29 10:57:44.953820] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 [2024-10-29 10:57:44.965746] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.552 [2024-10-29 10:57:44.965810] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.552 [2024-10-29 10:57:44.965820] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.552 [2024-10-29 10:57:44.965829] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.552 [2024-10-29 10:57:44.965835] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:39.552 [2024-10-29 10:57:44.965845] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 [2024-10-29 10:57:44.992969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:39.552 BaseBdev1 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.552 10:57:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.552 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:39.552 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.552 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.552 [ 00:08:39.552 { 00:08:39.552 "name": "BaseBdev1", 00:08:39.552 "aliases": [ 00:08:39.552 "5a977629-a3e2-43b0-9df5-8f3bfe214029" 00:08:39.552 ], 00:08:39.552 "product_name": "Malloc disk", 00:08:39.552 "block_size": 512, 00:08:39.552 "num_blocks": 65536, 00:08:39.552 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:39.552 "assigned_rate_limits": { 00:08:39.552 "rw_ios_per_sec": 0, 00:08:39.552 "rw_mbytes_per_sec": 0, 00:08:39.552 "r_mbytes_per_sec": 0, 00:08:39.552 "w_mbytes_per_sec": 0 00:08:39.552 }, 00:08:39.552 "claimed": true, 00:08:39.552 "claim_type": "exclusive_write", 00:08:39.552 "zoned": false, 00:08:39.552 "supported_io_types": { 00:08:39.552 "read": true, 00:08:39.552 "write": true, 00:08:39.552 "unmap": true, 00:08:39.552 "flush": true, 00:08:39.552 "reset": true, 00:08:39.552 "nvme_admin": false, 00:08:39.552 "nvme_io": false, 00:08:39.552 "nvme_io_md": false, 00:08:39.552 "write_zeroes": true, 00:08:39.552 "zcopy": true, 00:08:39.552 "get_zone_info": false, 00:08:39.552 "zone_management": false, 00:08:39.553 "zone_append": false, 00:08:39.553 "compare": false, 00:08:39.553 "compare_and_write": false, 00:08:39.553 "abort": true, 00:08:39.553 "seek_hole": false, 00:08:39.553 "seek_data": false, 00:08:39.553 "copy": true, 00:08:39.553 "nvme_iov_md": false 00:08:39.553 }, 00:08:39.553 "memory_domains": [ 00:08:39.553 { 00:08:39.553 "dma_device_id": "system", 00:08:39.553 "dma_device_type": 1 00:08:39.553 }, 00:08:39.553 { 00:08:39.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.553 "dma_device_type": 2 00:08:39.553 } 00:08:39.553 ], 00:08:39.553 "driver_specific": {} 00:08:39.553 } 00:08:39.553 ] 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.553 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.812 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.812 "name": "Existed_Raid", 00:08:39.812 "uuid": "d3245b92-ffc5-41fe-88e7-3adc18fb7a4f", 00:08:39.812 "strip_size_kb": 64, 00:08:39.813 "state": "configuring", 00:08:39.813 "raid_level": "raid0", 00:08:39.813 "superblock": true, 00:08:39.813 "num_base_bdevs": 3, 00:08:39.813 "num_base_bdevs_discovered": 1, 00:08:39.813 "num_base_bdevs_operational": 3, 00:08:39.813 "base_bdevs_list": [ 00:08:39.813 { 00:08:39.813 "name": "BaseBdev1", 00:08:39.813 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:39.813 "is_configured": true, 00:08:39.813 "data_offset": 2048, 00:08:39.813 "data_size": 63488 00:08:39.813 }, 00:08:39.813 { 00:08:39.813 "name": "BaseBdev2", 00:08:39.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.813 "is_configured": false, 00:08:39.813 "data_offset": 0, 00:08:39.813 "data_size": 0 00:08:39.813 }, 00:08:39.813 { 00:08:39.813 "name": "BaseBdev3", 00:08:39.813 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.813 "is_configured": false, 00:08:39.813 "data_offset": 0, 00:08:39.813 "data_size": 0 00:08:39.813 } 00:08:39.813 ] 00:08:39.813 }' 00:08:39.813 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.813 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.072 [2024-10-29 10:57:45.380401] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.072 [2024-10-29 10:57:45.380482] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.072 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.072 [2024-10-29 10:57:45.392382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.073 [2024-10-29 10:57:45.394557] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.073 [2024-10-29 10:57:45.394594] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.073 [2024-10-29 10:57:45.394604] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:40.073 [2024-10-29 10:57:45.394615] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.073 "name": "Existed_Raid", 00:08:40.073 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:40.073 "strip_size_kb": 64, 00:08:40.073 "state": "configuring", 00:08:40.073 "raid_level": "raid0", 00:08:40.073 "superblock": true, 00:08:40.073 "num_base_bdevs": 3, 00:08:40.073 "num_base_bdevs_discovered": 1, 00:08:40.073 "num_base_bdevs_operational": 3, 00:08:40.073 "base_bdevs_list": [ 00:08:40.073 { 00:08:40.073 "name": "BaseBdev1", 00:08:40.073 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:40.073 "is_configured": true, 00:08:40.073 "data_offset": 2048, 00:08:40.073 "data_size": 63488 00:08:40.073 }, 00:08:40.073 { 00:08:40.073 "name": "BaseBdev2", 00:08:40.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.073 "is_configured": false, 00:08:40.073 "data_offset": 0, 00:08:40.073 "data_size": 0 00:08:40.073 }, 00:08:40.073 { 00:08:40.073 "name": "BaseBdev3", 00:08:40.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.073 "is_configured": false, 00:08:40.073 "data_offset": 0, 00:08:40.073 "data_size": 0 00:08:40.073 } 00:08:40.073 ] 00:08:40.073 }' 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.073 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.333 [2024-10-29 10:57:45.724644] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:40.333 BaseBdev2 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.333 [ 00:08:40.333 { 00:08:40.333 "name": "BaseBdev2", 00:08:40.333 "aliases": [ 00:08:40.333 "eb676774-7b71-4ff4-9a6f-49c22ed839b8" 00:08:40.333 ], 00:08:40.333 "product_name": "Malloc disk", 00:08:40.333 "block_size": 512, 00:08:40.333 "num_blocks": 65536, 00:08:40.333 "uuid": "eb676774-7b71-4ff4-9a6f-49c22ed839b8", 00:08:40.333 "assigned_rate_limits": { 00:08:40.333 "rw_ios_per_sec": 0, 00:08:40.333 "rw_mbytes_per_sec": 0, 00:08:40.333 "r_mbytes_per_sec": 0, 00:08:40.333 "w_mbytes_per_sec": 0 00:08:40.333 }, 00:08:40.333 "claimed": true, 00:08:40.333 "claim_type": "exclusive_write", 00:08:40.333 "zoned": false, 00:08:40.333 "supported_io_types": { 00:08:40.333 "read": true, 00:08:40.333 "write": true, 00:08:40.333 "unmap": true, 00:08:40.333 "flush": true, 00:08:40.333 "reset": true, 00:08:40.333 "nvme_admin": false, 00:08:40.333 "nvme_io": false, 00:08:40.333 "nvme_io_md": false, 00:08:40.333 "write_zeroes": true, 00:08:40.333 "zcopy": true, 00:08:40.333 "get_zone_info": false, 00:08:40.333 "zone_management": false, 00:08:40.333 "zone_append": false, 00:08:40.333 "compare": false, 00:08:40.333 "compare_and_write": false, 00:08:40.333 "abort": true, 00:08:40.333 "seek_hole": false, 00:08:40.333 "seek_data": false, 00:08:40.333 "copy": true, 00:08:40.333 "nvme_iov_md": false 00:08:40.333 }, 00:08:40.333 "memory_domains": [ 00:08:40.333 { 00:08:40.333 "dma_device_id": "system", 00:08:40.333 "dma_device_type": 1 00:08:40.333 }, 00:08:40.333 { 00:08:40.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.333 "dma_device_type": 2 00:08:40.333 } 00:08:40.333 ], 00:08:40.333 "driver_specific": {} 00:08:40.333 } 00:08:40.333 ] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.333 "name": "Existed_Raid", 00:08:40.333 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:40.333 "strip_size_kb": 64, 00:08:40.333 "state": "configuring", 00:08:40.333 "raid_level": "raid0", 00:08:40.333 "superblock": true, 00:08:40.333 "num_base_bdevs": 3, 00:08:40.333 "num_base_bdevs_discovered": 2, 00:08:40.333 "num_base_bdevs_operational": 3, 00:08:40.333 "base_bdevs_list": [ 00:08:40.333 { 00:08:40.333 "name": "BaseBdev1", 00:08:40.333 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:40.333 "is_configured": true, 00:08:40.333 "data_offset": 2048, 00:08:40.333 "data_size": 63488 00:08:40.333 }, 00:08:40.333 { 00:08:40.333 "name": "BaseBdev2", 00:08:40.333 "uuid": "eb676774-7b71-4ff4-9a6f-49c22ed839b8", 00:08:40.333 "is_configured": true, 00:08:40.333 "data_offset": 2048, 00:08:40.333 "data_size": 63488 00:08:40.333 }, 00:08:40.333 { 00:08:40.333 "name": "BaseBdev3", 00:08:40.333 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.333 "is_configured": false, 00:08:40.333 "data_offset": 0, 00:08:40.333 "data_size": 0 00:08:40.333 } 00:08:40.333 ] 00:08:40.333 }' 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.333 10:57:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.902 [2024-10-29 10:57:46.229576] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:40.902 [2024-10-29 10:57:46.229809] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:40.902 [2024-10-29 10:57:46.229843] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:40.902 [2024-10-29 10:57:46.230186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:40.902 BaseBdev3 00:08:40.902 [2024-10-29 10:57:46.230387] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:40.902 [2024-10-29 10:57:46.230406] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:08:40.902 [2024-10-29 10:57:46.230593] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.902 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.902 [ 00:08:40.902 { 00:08:40.902 "name": "BaseBdev3", 00:08:40.902 "aliases": [ 00:08:40.902 "e6631065-1f7d-48c5-998d-2a75a478f35c" 00:08:40.902 ], 00:08:40.902 "product_name": "Malloc disk", 00:08:40.902 "block_size": 512, 00:08:40.902 "num_blocks": 65536, 00:08:40.902 "uuid": "e6631065-1f7d-48c5-998d-2a75a478f35c", 00:08:40.902 "assigned_rate_limits": { 00:08:40.902 "rw_ios_per_sec": 0, 00:08:40.902 "rw_mbytes_per_sec": 0, 00:08:40.902 "r_mbytes_per_sec": 0, 00:08:40.902 "w_mbytes_per_sec": 0 00:08:40.902 }, 00:08:40.902 "claimed": true, 00:08:40.902 "claim_type": "exclusive_write", 00:08:40.902 "zoned": false, 00:08:40.902 "supported_io_types": { 00:08:40.902 "read": true, 00:08:40.902 "write": true, 00:08:40.902 "unmap": true, 00:08:40.902 "flush": true, 00:08:40.902 "reset": true, 00:08:40.902 "nvme_admin": false, 00:08:40.902 "nvme_io": false, 00:08:40.902 "nvme_io_md": false, 00:08:40.902 "write_zeroes": true, 00:08:40.902 "zcopy": true, 00:08:40.902 "get_zone_info": false, 00:08:40.902 "zone_management": false, 00:08:40.902 "zone_append": false, 00:08:40.902 "compare": false, 00:08:40.902 "compare_and_write": false, 00:08:40.902 "abort": true, 00:08:40.902 "seek_hole": false, 00:08:40.902 "seek_data": false, 00:08:40.902 "copy": true, 00:08:40.903 "nvme_iov_md": false 00:08:40.903 }, 00:08:40.903 "memory_domains": [ 00:08:40.903 { 00:08:40.903 "dma_device_id": "system", 00:08:40.903 "dma_device_type": 1 00:08:40.903 }, 00:08:40.903 { 00:08:40.903 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.903 "dma_device_type": 2 00:08:40.903 } 00:08:40.903 ], 00:08:40.903 "driver_specific": {} 00:08:40.903 } 00:08:40.903 ] 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.903 "name": "Existed_Raid", 00:08:40.903 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:40.903 "strip_size_kb": 64, 00:08:40.903 "state": "online", 00:08:40.903 "raid_level": "raid0", 00:08:40.903 "superblock": true, 00:08:40.903 "num_base_bdevs": 3, 00:08:40.903 "num_base_bdevs_discovered": 3, 00:08:40.903 "num_base_bdevs_operational": 3, 00:08:40.903 "base_bdevs_list": [ 00:08:40.903 { 00:08:40.903 "name": "BaseBdev1", 00:08:40.903 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:40.903 "is_configured": true, 00:08:40.903 "data_offset": 2048, 00:08:40.903 "data_size": 63488 00:08:40.903 }, 00:08:40.903 { 00:08:40.903 "name": "BaseBdev2", 00:08:40.903 "uuid": "eb676774-7b71-4ff4-9a6f-49c22ed839b8", 00:08:40.903 "is_configured": true, 00:08:40.903 "data_offset": 2048, 00:08:40.903 "data_size": 63488 00:08:40.903 }, 00:08:40.903 { 00:08:40.903 "name": "BaseBdev3", 00:08:40.903 "uuid": "e6631065-1f7d-48c5-998d-2a75a478f35c", 00:08:40.903 "is_configured": true, 00:08:40.903 "data_offset": 2048, 00:08:40.903 "data_size": 63488 00:08:40.903 } 00:08:40.903 ] 00:08:40.903 }' 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.903 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:41.480 [2024-10-29 10:57:46.721159] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:41.480 "name": "Existed_Raid", 00:08:41.480 "aliases": [ 00:08:41.480 "c0d218db-df9a-4c61-b5eb-47db3895b55c" 00:08:41.480 ], 00:08:41.480 "product_name": "Raid Volume", 00:08:41.480 "block_size": 512, 00:08:41.480 "num_blocks": 190464, 00:08:41.480 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:41.480 "assigned_rate_limits": { 00:08:41.480 "rw_ios_per_sec": 0, 00:08:41.480 "rw_mbytes_per_sec": 0, 00:08:41.480 "r_mbytes_per_sec": 0, 00:08:41.480 "w_mbytes_per_sec": 0 00:08:41.480 }, 00:08:41.480 "claimed": false, 00:08:41.480 "zoned": false, 00:08:41.480 "supported_io_types": { 00:08:41.480 "read": true, 00:08:41.480 "write": true, 00:08:41.480 "unmap": true, 00:08:41.480 "flush": true, 00:08:41.480 "reset": true, 00:08:41.480 "nvme_admin": false, 00:08:41.480 "nvme_io": false, 00:08:41.480 "nvme_io_md": false, 00:08:41.480 "write_zeroes": true, 00:08:41.480 "zcopy": false, 00:08:41.480 "get_zone_info": false, 00:08:41.480 "zone_management": false, 00:08:41.480 "zone_append": false, 00:08:41.480 "compare": false, 00:08:41.480 "compare_and_write": false, 00:08:41.480 "abort": false, 00:08:41.480 "seek_hole": false, 00:08:41.480 "seek_data": false, 00:08:41.480 "copy": false, 00:08:41.480 "nvme_iov_md": false 00:08:41.480 }, 00:08:41.480 "memory_domains": [ 00:08:41.480 { 00:08:41.480 "dma_device_id": "system", 00:08:41.480 "dma_device_type": 1 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.480 "dma_device_type": 2 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "dma_device_id": "system", 00:08:41.480 "dma_device_type": 1 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.480 "dma_device_type": 2 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "dma_device_id": "system", 00:08:41.480 "dma_device_type": 1 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.480 "dma_device_type": 2 00:08:41.480 } 00:08:41.480 ], 00:08:41.480 "driver_specific": { 00:08:41.480 "raid": { 00:08:41.480 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:41.480 "strip_size_kb": 64, 00:08:41.480 "state": "online", 00:08:41.480 "raid_level": "raid0", 00:08:41.480 "superblock": true, 00:08:41.480 "num_base_bdevs": 3, 00:08:41.480 "num_base_bdevs_discovered": 3, 00:08:41.480 "num_base_bdevs_operational": 3, 00:08:41.480 "base_bdevs_list": [ 00:08:41.480 { 00:08:41.480 "name": "BaseBdev1", 00:08:41.480 "uuid": "5a977629-a3e2-43b0-9df5-8f3bfe214029", 00:08:41.480 "is_configured": true, 00:08:41.480 "data_offset": 2048, 00:08:41.480 "data_size": 63488 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "name": "BaseBdev2", 00:08:41.480 "uuid": "eb676774-7b71-4ff4-9a6f-49c22ed839b8", 00:08:41.480 "is_configured": true, 00:08:41.480 "data_offset": 2048, 00:08:41.480 "data_size": 63488 00:08:41.480 }, 00:08:41.480 { 00:08:41.480 "name": "BaseBdev3", 00:08:41.480 "uuid": "e6631065-1f7d-48c5-998d-2a75a478f35c", 00:08:41.480 "is_configured": true, 00:08:41.480 "data_offset": 2048, 00:08:41.480 "data_size": 63488 00:08:41.480 } 00:08:41.480 ] 00:08:41.480 } 00:08:41.480 } 00:08:41.480 }' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:41.480 BaseBdev2 00:08:41.480 BaseBdev3' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.480 [2024-10-29 10:57:46.956555] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:41.480 [2024-10-29 10:57:46.956602] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:41.480 [2024-10-29 10:57:46.956682] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:41.480 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.740 10:57:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.741 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.741 10:57:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.741 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.741 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.741 "name": "Existed_Raid", 00:08:41.741 "uuid": "c0d218db-df9a-4c61-b5eb-47db3895b55c", 00:08:41.741 "strip_size_kb": 64, 00:08:41.741 "state": "offline", 00:08:41.741 "raid_level": "raid0", 00:08:41.741 "superblock": true, 00:08:41.741 "num_base_bdevs": 3, 00:08:41.741 "num_base_bdevs_discovered": 2, 00:08:41.741 "num_base_bdevs_operational": 2, 00:08:41.741 "base_bdevs_list": [ 00:08:41.741 { 00:08:41.741 "name": null, 00:08:41.741 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.741 "is_configured": false, 00:08:41.741 "data_offset": 0, 00:08:41.741 "data_size": 63488 00:08:41.741 }, 00:08:41.741 { 00:08:41.741 "name": "BaseBdev2", 00:08:41.741 "uuid": "eb676774-7b71-4ff4-9a6f-49c22ed839b8", 00:08:41.741 "is_configured": true, 00:08:41.741 "data_offset": 2048, 00:08:41.741 "data_size": 63488 00:08:41.741 }, 00:08:41.741 { 00:08:41.741 "name": "BaseBdev3", 00:08:41.741 "uuid": "e6631065-1f7d-48c5-998d-2a75a478f35c", 00:08:41.741 "is_configured": true, 00:08:41.741 "data_offset": 2048, 00:08:41.741 "data_size": 63488 00:08:41.741 } 00:08:41.741 ] 00:08:41.741 }' 00:08:41.741 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.741 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.001 [2024-10-29 10:57:47.476780] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:42.001 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.261 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 [2024-10-29 10:57:47.553716] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:42.262 [2024-10-29 10:57:47.553789] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 BaseBdev2 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 [ 00:08:42.262 { 00:08:42.262 "name": "BaseBdev2", 00:08:42.262 "aliases": [ 00:08:42.262 "564a5cda-9953-47aa-ad22-bebaf36425a3" 00:08:42.262 ], 00:08:42.262 "product_name": "Malloc disk", 00:08:42.262 "block_size": 512, 00:08:42.262 "num_blocks": 65536, 00:08:42.262 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:42.262 "assigned_rate_limits": { 00:08:42.262 "rw_ios_per_sec": 0, 00:08:42.262 "rw_mbytes_per_sec": 0, 00:08:42.262 "r_mbytes_per_sec": 0, 00:08:42.262 "w_mbytes_per_sec": 0 00:08:42.262 }, 00:08:42.262 "claimed": false, 00:08:42.262 "zoned": false, 00:08:42.262 "supported_io_types": { 00:08:42.262 "read": true, 00:08:42.262 "write": true, 00:08:42.262 "unmap": true, 00:08:42.262 "flush": true, 00:08:42.262 "reset": true, 00:08:42.262 "nvme_admin": false, 00:08:42.262 "nvme_io": false, 00:08:42.262 "nvme_io_md": false, 00:08:42.262 "write_zeroes": true, 00:08:42.262 "zcopy": true, 00:08:42.262 "get_zone_info": false, 00:08:42.262 "zone_management": false, 00:08:42.262 "zone_append": false, 00:08:42.262 "compare": false, 00:08:42.262 "compare_and_write": false, 00:08:42.262 "abort": true, 00:08:42.262 "seek_hole": false, 00:08:42.262 "seek_data": false, 00:08:42.262 "copy": true, 00:08:42.262 "nvme_iov_md": false 00:08:42.262 }, 00:08:42.262 "memory_domains": [ 00:08:42.262 { 00:08:42.262 "dma_device_id": "system", 00:08:42.262 "dma_device_type": 1 00:08:42.262 }, 00:08:42.262 { 00:08:42.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.262 "dma_device_type": 2 00:08:42.262 } 00:08:42.262 ], 00:08:42.262 "driver_specific": {} 00:08:42.262 } 00:08:42.262 ] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 BaseBdev3 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 [ 00:08:42.262 { 00:08:42.262 "name": "BaseBdev3", 00:08:42.262 "aliases": [ 00:08:42.262 "bd02ccb5-25e1-4e77-9379-0f59964023e3" 00:08:42.262 ], 00:08:42.262 "product_name": "Malloc disk", 00:08:42.262 "block_size": 512, 00:08:42.262 "num_blocks": 65536, 00:08:42.262 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:42.262 "assigned_rate_limits": { 00:08:42.262 "rw_ios_per_sec": 0, 00:08:42.262 "rw_mbytes_per_sec": 0, 00:08:42.262 "r_mbytes_per_sec": 0, 00:08:42.262 "w_mbytes_per_sec": 0 00:08:42.262 }, 00:08:42.262 "claimed": false, 00:08:42.262 "zoned": false, 00:08:42.262 "supported_io_types": { 00:08:42.262 "read": true, 00:08:42.262 "write": true, 00:08:42.262 "unmap": true, 00:08:42.262 "flush": true, 00:08:42.262 "reset": true, 00:08:42.262 "nvme_admin": false, 00:08:42.262 "nvme_io": false, 00:08:42.262 "nvme_io_md": false, 00:08:42.262 "write_zeroes": true, 00:08:42.262 "zcopy": true, 00:08:42.262 "get_zone_info": false, 00:08:42.262 "zone_management": false, 00:08:42.262 "zone_append": false, 00:08:42.262 "compare": false, 00:08:42.262 "compare_and_write": false, 00:08:42.262 "abort": true, 00:08:42.262 "seek_hole": false, 00:08:42.262 "seek_data": false, 00:08:42.262 "copy": true, 00:08:42.262 "nvme_iov_md": false 00:08:42.262 }, 00:08:42.262 "memory_domains": [ 00:08:42.262 { 00:08:42.262 "dma_device_id": "system", 00:08:42.262 "dma_device_type": 1 00:08:42.262 }, 00:08:42.262 { 00:08:42.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.262 "dma_device_type": 2 00:08:42.262 } 00:08:42.262 ], 00:08:42.262 "driver_specific": {} 00:08:42.262 } 00:08:42.262 ] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.262 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.262 [2024-10-29 10:57:47.744860] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:42.263 [2024-10-29 10:57:47.744930] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:42.263 [2024-10-29 10:57:47.744963] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:42.263 [2024-10-29 10:57:47.747229] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.263 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.521 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.521 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.521 "name": "Existed_Raid", 00:08:42.521 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:42.521 "strip_size_kb": 64, 00:08:42.521 "state": "configuring", 00:08:42.521 "raid_level": "raid0", 00:08:42.521 "superblock": true, 00:08:42.521 "num_base_bdevs": 3, 00:08:42.521 "num_base_bdevs_discovered": 2, 00:08:42.521 "num_base_bdevs_operational": 3, 00:08:42.521 "base_bdevs_list": [ 00:08:42.521 { 00:08:42.521 "name": "BaseBdev1", 00:08:42.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.521 "is_configured": false, 00:08:42.521 "data_offset": 0, 00:08:42.521 "data_size": 0 00:08:42.521 }, 00:08:42.521 { 00:08:42.521 "name": "BaseBdev2", 00:08:42.521 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:42.521 "is_configured": true, 00:08:42.521 "data_offset": 2048, 00:08:42.521 "data_size": 63488 00:08:42.521 }, 00:08:42.521 { 00:08:42.521 "name": "BaseBdev3", 00:08:42.521 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:42.521 "is_configured": true, 00:08:42.521 "data_offset": 2048, 00:08:42.521 "data_size": 63488 00:08:42.521 } 00:08:42.521 ] 00:08:42.521 }' 00:08:42.521 10:57:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.521 10:57:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.781 [2024-10-29 10:57:48.148254] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.781 "name": "Existed_Raid", 00:08:42.781 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:42.781 "strip_size_kb": 64, 00:08:42.781 "state": "configuring", 00:08:42.781 "raid_level": "raid0", 00:08:42.781 "superblock": true, 00:08:42.781 "num_base_bdevs": 3, 00:08:42.781 "num_base_bdevs_discovered": 1, 00:08:42.781 "num_base_bdevs_operational": 3, 00:08:42.781 "base_bdevs_list": [ 00:08:42.781 { 00:08:42.781 "name": "BaseBdev1", 00:08:42.781 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.781 "is_configured": false, 00:08:42.781 "data_offset": 0, 00:08:42.781 "data_size": 0 00:08:42.781 }, 00:08:42.781 { 00:08:42.781 "name": null, 00:08:42.781 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:42.781 "is_configured": false, 00:08:42.781 "data_offset": 0, 00:08:42.781 "data_size": 63488 00:08:42.781 }, 00:08:42.781 { 00:08:42.781 "name": "BaseBdev3", 00:08:42.781 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:42.781 "is_configured": true, 00:08:42.781 "data_offset": 2048, 00:08:42.781 "data_size": 63488 00:08:42.781 } 00:08:42.781 ] 00:08:42.781 }' 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.781 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.351 [2024-10-29 10:57:48.628323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:43.351 BaseBdev1 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.351 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.351 [ 00:08:43.351 { 00:08:43.351 "name": "BaseBdev1", 00:08:43.351 "aliases": [ 00:08:43.351 "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a" 00:08:43.351 ], 00:08:43.351 "product_name": "Malloc disk", 00:08:43.351 "block_size": 512, 00:08:43.351 "num_blocks": 65536, 00:08:43.351 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:43.351 "assigned_rate_limits": { 00:08:43.351 "rw_ios_per_sec": 0, 00:08:43.351 "rw_mbytes_per_sec": 0, 00:08:43.351 "r_mbytes_per_sec": 0, 00:08:43.351 "w_mbytes_per_sec": 0 00:08:43.351 }, 00:08:43.351 "claimed": true, 00:08:43.351 "claim_type": "exclusive_write", 00:08:43.351 "zoned": false, 00:08:43.351 "supported_io_types": { 00:08:43.352 "read": true, 00:08:43.352 "write": true, 00:08:43.352 "unmap": true, 00:08:43.352 "flush": true, 00:08:43.352 "reset": true, 00:08:43.352 "nvme_admin": false, 00:08:43.352 "nvme_io": false, 00:08:43.352 "nvme_io_md": false, 00:08:43.352 "write_zeroes": true, 00:08:43.352 "zcopy": true, 00:08:43.352 "get_zone_info": false, 00:08:43.352 "zone_management": false, 00:08:43.352 "zone_append": false, 00:08:43.352 "compare": false, 00:08:43.352 "compare_and_write": false, 00:08:43.352 "abort": true, 00:08:43.352 "seek_hole": false, 00:08:43.352 "seek_data": false, 00:08:43.352 "copy": true, 00:08:43.352 "nvme_iov_md": false 00:08:43.352 }, 00:08:43.352 "memory_domains": [ 00:08:43.352 { 00:08:43.352 "dma_device_id": "system", 00:08:43.352 "dma_device_type": 1 00:08:43.352 }, 00:08:43.352 { 00:08:43.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.352 "dma_device_type": 2 00:08:43.352 } 00:08:43.352 ], 00:08:43.352 "driver_specific": {} 00:08:43.352 } 00:08:43.352 ] 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.352 "name": "Existed_Raid", 00:08:43.352 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:43.352 "strip_size_kb": 64, 00:08:43.352 "state": "configuring", 00:08:43.352 "raid_level": "raid0", 00:08:43.352 "superblock": true, 00:08:43.352 "num_base_bdevs": 3, 00:08:43.352 "num_base_bdevs_discovered": 2, 00:08:43.352 "num_base_bdevs_operational": 3, 00:08:43.352 "base_bdevs_list": [ 00:08:43.352 { 00:08:43.352 "name": "BaseBdev1", 00:08:43.352 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:43.352 "is_configured": true, 00:08:43.352 "data_offset": 2048, 00:08:43.352 "data_size": 63488 00:08:43.352 }, 00:08:43.352 { 00:08:43.352 "name": null, 00:08:43.352 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:43.352 "is_configured": false, 00:08:43.352 "data_offset": 0, 00:08:43.352 "data_size": 63488 00:08:43.352 }, 00:08:43.352 { 00:08:43.352 "name": "BaseBdev3", 00:08:43.352 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:43.352 "is_configured": true, 00:08:43.352 "data_offset": 2048, 00:08:43.352 "data_size": 63488 00:08:43.352 } 00:08:43.352 ] 00:08:43.352 }' 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.352 10:57:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.612 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:43.612 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.612 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.612 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.872 [2024-10-29 10:57:49.123573] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.872 "name": "Existed_Raid", 00:08:43.872 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:43.872 "strip_size_kb": 64, 00:08:43.872 "state": "configuring", 00:08:43.872 "raid_level": "raid0", 00:08:43.872 "superblock": true, 00:08:43.872 "num_base_bdevs": 3, 00:08:43.872 "num_base_bdevs_discovered": 1, 00:08:43.872 "num_base_bdevs_operational": 3, 00:08:43.872 "base_bdevs_list": [ 00:08:43.872 { 00:08:43.872 "name": "BaseBdev1", 00:08:43.872 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:43.872 "is_configured": true, 00:08:43.872 "data_offset": 2048, 00:08:43.872 "data_size": 63488 00:08:43.872 }, 00:08:43.872 { 00:08:43.872 "name": null, 00:08:43.872 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:43.872 "is_configured": false, 00:08:43.872 "data_offset": 0, 00:08:43.872 "data_size": 63488 00:08:43.872 }, 00:08:43.872 { 00:08:43.872 "name": null, 00:08:43.872 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:43.872 "is_configured": false, 00:08:43.872 "data_offset": 0, 00:08:43.872 "data_size": 63488 00:08:43.872 } 00:08:43.872 ] 00:08:43.872 }' 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.872 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.132 [2024-10-29 10:57:49.602834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.132 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.391 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.391 "name": "Existed_Raid", 00:08:44.391 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:44.391 "strip_size_kb": 64, 00:08:44.391 "state": "configuring", 00:08:44.391 "raid_level": "raid0", 00:08:44.391 "superblock": true, 00:08:44.391 "num_base_bdevs": 3, 00:08:44.391 "num_base_bdevs_discovered": 2, 00:08:44.391 "num_base_bdevs_operational": 3, 00:08:44.391 "base_bdevs_list": [ 00:08:44.391 { 00:08:44.391 "name": "BaseBdev1", 00:08:44.391 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:44.391 "is_configured": true, 00:08:44.391 "data_offset": 2048, 00:08:44.391 "data_size": 63488 00:08:44.391 }, 00:08:44.391 { 00:08:44.391 "name": null, 00:08:44.391 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:44.391 "is_configured": false, 00:08:44.391 "data_offset": 0, 00:08:44.391 "data_size": 63488 00:08:44.391 }, 00:08:44.391 { 00:08:44.391 "name": "BaseBdev3", 00:08:44.391 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:44.391 "is_configured": true, 00:08:44.391 "data_offset": 2048, 00:08:44.391 "data_size": 63488 00:08:44.391 } 00:08:44.391 ] 00:08:44.391 }' 00:08:44.391 10:57:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.391 10:57:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.650 [2024-10-29 10:57:50.066079] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.650 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.651 "name": "Existed_Raid", 00:08:44.651 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:44.651 "strip_size_kb": 64, 00:08:44.651 "state": "configuring", 00:08:44.651 "raid_level": "raid0", 00:08:44.651 "superblock": true, 00:08:44.651 "num_base_bdevs": 3, 00:08:44.651 "num_base_bdevs_discovered": 1, 00:08:44.651 "num_base_bdevs_operational": 3, 00:08:44.651 "base_bdevs_list": [ 00:08:44.651 { 00:08:44.651 "name": null, 00:08:44.651 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:44.651 "is_configured": false, 00:08:44.651 "data_offset": 0, 00:08:44.651 "data_size": 63488 00:08:44.651 }, 00:08:44.651 { 00:08:44.651 "name": null, 00:08:44.651 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:44.651 "is_configured": false, 00:08:44.651 "data_offset": 0, 00:08:44.651 "data_size": 63488 00:08:44.651 }, 00:08:44.651 { 00:08:44.651 "name": "BaseBdev3", 00:08:44.651 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:44.651 "is_configured": true, 00:08:44.651 "data_offset": 2048, 00:08:44.651 "data_size": 63488 00:08:44.651 } 00:08:44.651 ] 00:08:44.651 }' 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.651 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.221 [2024-10-29 10:57:50.525254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.221 "name": "Existed_Raid", 00:08:45.221 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:45.221 "strip_size_kb": 64, 00:08:45.221 "state": "configuring", 00:08:45.221 "raid_level": "raid0", 00:08:45.221 "superblock": true, 00:08:45.221 "num_base_bdevs": 3, 00:08:45.221 "num_base_bdevs_discovered": 2, 00:08:45.221 "num_base_bdevs_operational": 3, 00:08:45.221 "base_bdevs_list": [ 00:08:45.221 { 00:08:45.221 "name": null, 00:08:45.221 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:45.221 "is_configured": false, 00:08:45.221 "data_offset": 0, 00:08:45.221 "data_size": 63488 00:08:45.221 }, 00:08:45.221 { 00:08:45.221 "name": "BaseBdev2", 00:08:45.221 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:45.221 "is_configured": true, 00:08:45.221 "data_offset": 2048, 00:08:45.221 "data_size": 63488 00:08:45.221 }, 00:08:45.221 { 00:08:45.221 "name": "BaseBdev3", 00:08:45.221 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:45.221 "is_configured": true, 00:08:45.221 "data_offset": 2048, 00:08:45.221 "data_size": 63488 00:08:45.221 } 00:08:45.221 ] 00:08:45.221 }' 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.221 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.481 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:45.481 10:57:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.481 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.481 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.741 10:57:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.741 [2024-10-29 10:57:51.057430] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:45.741 [2024-10-29 10:57:51.057625] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:45.741 [2024-10-29 10:57:51.057643] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:45.741 [2024-10-29 10:57:51.057985] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:08:45.741 [2024-10-29 10:57:51.058122] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:45.741 [2024-10-29 10:57:51.058135] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:08:45.741 NewBaseBdev 00:08:45.741 [2024-10-29 10:57:51.058290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.741 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.742 [ 00:08:45.742 { 00:08:45.742 "name": "NewBaseBdev", 00:08:45.742 "aliases": [ 00:08:45.742 "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a" 00:08:45.742 ], 00:08:45.742 "product_name": "Malloc disk", 00:08:45.742 "block_size": 512, 00:08:45.742 "num_blocks": 65536, 00:08:45.742 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:45.742 "assigned_rate_limits": { 00:08:45.742 "rw_ios_per_sec": 0, 00:08:45.742 "rw_mbytes_per_sec": 0, 00:08:45.742 "r_mbytes_per_sec": 0, 00:08:45.742 "w_mbytes_per_sec": 0 00:08:45.742 }, 00:08:45.742 "claimed": true, 00:08:45.742 "claim_type": "exclusive_write", 00:08:45.742 "zoned": false, 00:08:45.742 "supported_io_types": { 00:08:45.742 "read": true, 00:08:45.742 "write": true, 00:08:45.742 "unmap": true, 00:08:45.742 "flush": true, 00:08:45.742 "reset": true, 00:08:45.742 "nvme_admin": false, 00:08:45.742 "nvme_io": false, 00:08:45.742 "nvme_io_md": false, 00:08:45.742 "write_zeroes": true, 00:08:45.742 "zcopy": true, 00:08:45.742 "get_zone_info": false, 00:08:45.742 "zone_management": false, 00:08:45.742 "zone_append": false, 00:08:45.742 "compare": false, 00:08:45.742 "compare_and_write": false, 00:08:45.742 "abort": true, 00:08:45.742 "seek_hole": false, 00:08:45.742 "seek_data": false, 00:08:45.742 "copy": true, 00:08:45.742 "nvme_iov_md": false 00:08:45.742 }, 00:08:45.742 "memory_domains": [ 00:08:45.742 { 00:08:45.742 "dma_device_id": "system", 00:08:45.742 "dma_device_type": 1 00:08:45.742 }, 00:08:45.742 { 00:08:45.742 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:45.742 "dma_device_type": 2 00:08:45.742 } 00:08:45.742 ], 00:08:45.742 "driver_specific": {} 00:08:45.742 } 00:08:45.742 ] 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.742 "name": "Existed_Raid", 00:08:45.742 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:45.742 "strip_size_kb": 64, 00:08:45.742 "state": "online", 00:08:45.742 "raid_level": "raid0", 00:08:45.742 "superblock": true, 00:08:45.742 "num_base_bdevs": 3, 00:08:45.742 "num_base_bdevs_discovered": 3, 00:08:45.742 "num_base_bdevs_operational": 3, 00:08:45.742 "base_bdevs_list": [ 00:08:45.742 { 00:08:45.742 "name": "NewBaseBdev", 00:08:45.742 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:45.742 "is_configured": true, 00:08:45.742 "data_offset": 2048, 00:08:45.742 "data_size": 63488 00:08:45.742 }, 00:08:45.742 { 00:08:45.742 "name": "BaseBdev2", 00:08:45.742 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:45.742 "is_configured": true, 00:08:45.742 "data_offset": 2048, 00:08:45.742 "data_size": 63488 00:08:45.742 }, 00:08:45.742 { 00:08:45.742 "name": "BaseBdev3", 00:08:45.742 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:45.742 "is_configured": true, 00:08:45.742 "data_offset": 2048, 00:08:45.742 "data_size": 63488 00:08:45.742 } 00:08:45.742 ] 00:08:45.742 }' 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.742 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:46.312 [2024-10-29 10:57:51.533040] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.312 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:46.312 "name": "Existed_Raid", 00:08:46.312 "aliases": [ 00:08:46.312 "864589ae-988e-45fd-9b26-192dfda20323" 00:08:46.312 ], 00:08:46.312 "product_name": "Raid Volume", 00:08:46.312 "block_size": 512, 00:08:46.312 "num_blocks": 190464, 00:08:46.312 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:46.312 "assigned_rate_limits": { 00:08:46.312 "rw_ios_per_sec": 0, 00:08:46.312 "rw_mbytes_per_sec": 0, 00:08:46.312 "r_mbytes_per_sec": 0, 00:08:46.312 "w_mbytes_per_sec": 0 00:08:46.312 }, 00:08:46.312 "claimed": false, 00:08:46.312 "zoned": false, 00:08:46.312 "supported_io_types": { 00:08:46.312 "read": true, 00:08:46.312 "write": true, 00:08:46.312 "unmap": true, 00:08:46.312 "flush": true, 00:08:46.312 "reset": true, 00:08:46.312 "nvme_admin": false, 00:08:46.312 "nvme_io": false, 00:08:46.312 "nvme_io_md": false, 00:08:46.312 "write_zeroes": true, 00:08:46.312 "zcopy": false, 00:08:46.312 "get_zone_info": false, 00:08:46.312 "zone_management": false, 00:08:46.312 "zone_append": false, 00:08:46.312 "compare": false, 00:08:46.312 "compare_and_write": false, 00:08:46.312 "abort": false, 00:08:46.312 "seek_hole": false, 00:08:46.312 "seek_data": false, 00:08:46.312 "copy": false, 00:08:46.312 "nvme_iov_md": false 00:08:46.312 }, 00:08:46.312 "memory_domains": [ 00:08:46.312 { 00:08:46.312 "dma_device_id": "system", 00:08:46.312 "dma_device_type": 1 00:08:46.312 }, 00:08:46.312 { 00:08:46.312 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.312 "dma_device_type": 2 00:08:46.312 }, 00:08:46.312 { 00:08:46.312 "dma_device_id": "system", 00:08:46.312 "dma_device_type": 1 00:08:46.312 }, 00:08:46.312 { 00:08:46.313 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.313 "dma_device_type": 2 00:08:46.313 }, 00:08:46.313 { 00:08:46.313 "dma_device_id": "system", 00:08:46.313 "dma_device_type": 1 00:08:46.313 }, 00:08:46.313 { 00:08:46.313 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.313 "dma_device_type": 2 00:08:46.313 } 00:08:46.313 ], 00:08:46.313 "driver_specific": { 00:08:46.313 "raid": { 00:08:46.313 "uuid": "864589ae-988e-45fd-9b26-192dfda20323", 00:08:46.313 "strip_size_kb": 64, 00:08:46.313 "state": "online", 00:08:46.313 "raid_level": "raid0", 00:08:46.313 "superblock": true, 00:08:46.313 "num_base_bdevs": 3, 00:08:46.313 "num_base_bdevs_discovered": 3, 00:08:46.313 "num_base_bdevs_operational": 3, 00:08:46.313 "base_bdevs_list": [ 00:08:46.313 { 00:08:46.313 "name": "NewBaseBdev", 00:08:46.313 "uuid": "ef880f3f-13af-4e0c-a972-e6e3d9ac9a3a", 00:08:46.313 "is_configured": true, 00:08:46.313 "data_offset": 2048, 00:08:46.313 "data_size": 63488 00:08:46.313 }, 00:08:46.313 { 00:08:46.313 "name": "BaseBdev2", 00:08:46.313 "uuid": "564a5cda-9953-47aa-ad22-bebaf36425a3", 00:08:46.313 "is_configured": true, 00:08:46.313 "data_offset": 2048, 00:08:46.313 "data_size": 63488 00:08:46.313 }, 00:08:46.313 { 00:08:46.313 "name": "BaseBdev3", 00:08:46.313 "uuid": "bd02ccb5-25e1-4e77-9379-0f59964023e3", 00:08:46.313 "is_configured": true, 00:08:46.313 "data_offset": 2048, 00:08:46.313 "data_size": 63488 00:08:46.313 } 00:08:46.313 ] 00:08:46.313 } 00:08:46.313 } 00:08:46.313 }' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:46.313 BaseBdev2 00:08:46.313 BaseBdev3' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.313 [2024-10-29 10:57:51.784230] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:46.313 [2024-10-29 10:57:51.784275] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:46.313 [2024-10-29 10:57:51.784371] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:46.313 [2024-10-29 10:57:51.784436] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:46.313 [2024-10-29 10:57:51.784450] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75780 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 75780 ']' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 75780 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:46.313 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 75780 00:08:46.573 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:46.573 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:46.573 killing process with pid 75780 00:08:46.573 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 75780' 00:08:46.573 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 75780 00:08:46.574 [2024-10-29 10:57:51.825246] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:46.574 10:57:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 75780 00:08:46.574 [2024-10-29 10:57:51.883367] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:46.834 10:57:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:46.834 00:08:46.834 real 0m8.545s 00:08:46.834 user 0m14.334s 00:08:46.834 sys 0m1.738s 00:08:46.834 10:57:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:46.834 10:57:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:46.834 ************************************ 00:08:46.834 END TEST raid_state_function_test_sb 00:08:46.834 ************************************ 00:08:46.834 10:57:52 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:46.834 10:57:52 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:08:46.834 10:57:52 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:46.834 10:57:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:46.834 ************************************ 00:08:46.834 START TEST raid_superblock_test 00:08:46.834 ************************************ 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid0 3 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=76384 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 76384 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 76384 ']' 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:46.834 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:46.834 10:57:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.094 [2024-10-29 10:57:52.374889] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:47.094 [2024-10-29 10:57:52.375040] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76384 ] 00:08:47.094 [2024-10-29 10:57:52.553081] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:47.094 [2024-10-29 10:57:52.593123] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.354 [2024-10-29 10:57:52.669718] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.354 [2024-10-29 10:57:52.669763] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 malloc1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 [2024-10-29 10:57:53.208988] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:47.925 [2024-10-29 10:57:53.209067] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.925 [2024-10-29 10:57:53.209090] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:47.925 [2024-10-29 10:57:53.209106] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.925 [2024-10-29 10:57:53.211560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.925 [2024-10-29 10:57:53.211602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:47.925 pt1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 malloc2 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 [2024-10-29 10:57:53.243674] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:47.925 [2024-10-29 10:57:53.243803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.925 [2024-10-29 10:57:53.243842] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:47.925 [2024-10-29 10:57:53.243875] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.925 [2024-10-29 10:57:53.246429] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.925 [2024-10-29 10:57:53.246505] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:47.925 pt2 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 malloc3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 [2024-10-29 10:57:53.282695] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:47.925 [2024-10-29 10:57:53.282812] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.925 [2024-10-29 10:57:53.282850] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:47.925 [2024-10-29 10:57:53.282881] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.925 [2024-10-29 10:57:53.285297] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.925 [2024-10-29 10:57:53.285376] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:47.925 pt3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.925 [2024-10-29 10:57:53.294708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:47.925 [2024-10-29 10:57:53.296868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:47.925 [2024-10-29 10:57:53.296972] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:47.925 [2024-10-29 10:57:53.297122] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:08:47.925 [2024-10-29 10:57:53.297138] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:47.925 [2024-10-29 10:57:53.297420] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:47.925 [2024-10-29 10:57:53.297560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:08:47.925 [2024-10-29 10:57:53.297578] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:08:47.925 [2024-10-29 10:57:53.297687] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.925 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.926 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.926 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.926 "name": "raid_bdev1", 00:08:47.926 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:47.926 "strip_size_kb": 64, 00:08:47.926 "state": "online", 00:08:47.926 "raid_level": "raid0", 00:08:47.926 "superblock": true, 00:08:47.926 "num_base_bdevs": 3, 00:08:47.926 "num_base_bdevs_discovered": 3, 00:08:47.926 "num_base_bdevs_operational": 3, 00:08:47.926 "base_bdevs_list": [ 00:08:47.926 { 00:08:47.926 "name": "pt1", 00:08:47.926 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:47.926 "is_configured": true, 00:08:47.926 "data_offset": 2048, 00:08:47.926 "data_size": 63488 00:08:47.926 }, 00:08:47.926 { 00:08:47.926 "name": "pt2", 00:08:47.926 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:47.926 "is_configured": true, 00:08:47.926 "data_offset": 2048, 00:08:47.926 "data_size": 63488 00:08:47.926 }, 00:08:47.926 { 00:08:47.926 "name": "pt3", 00:08:47.926 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:47.926 "is_configured": true, 00:08:47.926 "data_offset": 2048, 00:08:47.926 "data_size": 63488 00:08:47.926 } 00:08:47.926 ] 00:08:47.926 }' 00:08:47.926 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.926 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.496 [2024-10-29 10:57:53.750344] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.496 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:48.496 "name": "raid_bdev1", 00:08:48.496 "aliases": [ 00:08:48.496 "d6612ea0-5405-4cff-80c0-511cd015af02" 00:08:48.496 ], 00:08:48.496 "product_name": "Raid Volume", 00:08:48.496 "block_size": 512, 00:08:48.496 "num_blocks": 190464, 00:08:48.496 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:48.496 "assigned_rate_limits": { 00:08:48.496 "rw_ios_per_sec": 0, 00:08:48.496 "rw_mbytes_per_sec": 0, 00:08:48.496 "r_mbytes_per_sec": 0, 00:08:48.496 "w_mbytes_per_sec": 0 00:08:48.496 }, 00:08:48.496 "claimed": false, 00:08:48.496 "zoned": false, 00:08:48.496 "supported_io_types": { 00:08:48.496 "read": true, 00:08:48.496 "write": true, 00:08:48.496 "unmap": true, 00:08:48.496 "flush": true, 00:08:48.496 "reset": true, 00:08:48.496 "nvme_admin": false, 00:08:48.496 "nvme_io": false, 00:08:48.496 "nvme_io_md": false, 00:08:48.496 "write_zeroes": true, 00:08:48.496 "zcopy": false, 00:08:48.496 "get_zone_info": false, 00:08:48.496 "zone_management": false, 00:08:48.496 "zone_append": false, 00:08:48.496 "compare": false, 00:08:48.496 "compare_and_write": false, 00:08:48.496 "abort": false, 00:08:48.496 "seek_hole": false, 00:08:48.496 "seek_data": false, 00:08:48.496 "copy": false, 00:08:48.496 "nvme_iov_md": false 00:08:48.496 }, 00:08:48.496 "memory_domains": [ 00:08:48.496 { 00:08:48.496 "dma_device_id": "system", 00:08:48.496 "dma_device_type": 1 00:08:48.496 }, 00:08:48.496 { 00:08:48.496 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.496 "dma_device_type": 2 00:08:48.496 }, 00:08:48.496 { 00:08:48.496 "dma_device_id": "system", 00:08:48.496 "dma_device_type": 1 00:08:48.496 }, 00:08:48.496 { 00:08:48.496 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.496 "dma_device_type": 2 00:08:48.496 }, 00:08:48.496 { 00:08:48.496 "dma_device_id": "system", 00:08:48.496 "dma_device_type": 1 00:08:48.496 }, 00:08:48.496 { 00:08:48.496 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.496 "dma_device_type": 2 00:08:48.496 } 00:08:48.496 ], 00:08:48.496 "driver_specific": { 00:08:48.496 "raid": { 00:08:48.496 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:48.496 "strip_size_kb": 64, 00:08:48.496 "state": "online", 00:08:48.496 "raid_level": "raid0", 00:08:48.496 "superblock": true, 00:08:48.496 "num_base_bdevs": 3, 00:08:48.496 "num_base_bdevs_discovered": 3, 00:08:48.496 "num_base_bdevs_operational": 3, 00:08:48.496 "base_bdevs_list": [ 00:08:48.496 { 00:08:48.496 "name": "pt1", 00:08:48.497 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:48.497 "is_configured": true, 00:08:48.497 "data_offset": 2048, 00:08:48.497 "data_size": 63488 00:08:48.497 }, 00:08:48.497 { 00:08:48.497 "name": "pt2", 00:08:48.497 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:48.497 "is_configured": true, 00:08:48.497 "data_offset": 2048, 00:08:48.497 "data_size": 63488 00:08:48.497 }, 00:08:48.497 { 00:08:48.497 "name": "pt3", 00:08:48.497 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:48.497 "is_configured": true, 00:08:48.497 "data_offset": 2048, 00:08:48.497 "data_size": 63488 00:08:48.497 } 00:08:48.497 ] 00:08:48.497 } 00:08:48.497 } 00:08:48.497 }' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:48.497 pt2 00:08:48.497 pt3' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.497 [2024-10-29 10:57:53.953878] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d6612ea0-5405-4cff-80c0-511cd015af02 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z d6612ea0-5405-4cff-80c0-511cd015af02 ']' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.497 [2024-10-29 10:57:53.985530] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:48.497 [2024-10-29 10:57:53.985625] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:48.497 [2024-10-29 10:57:53.985755] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:48.497 [2024-10-29 10:57:53.985860] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:48.497 [2024-10-29 10:57:53.985906] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.497 10:57:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 [2024-10-29 10:57:54.133292] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:48.758 [2024-10-29 10:57:54.135680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:48.758 [2024-10-29 10:57:54.135771] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:48.758 [2024-10-29 10:57:54.135880] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:48.758 [2024-10-29 10:57:54.135998] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:48.758 [2024-10-29 10:57:54.136070] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:48.758 [2024-10-29 10:57:54.136150] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:48.758 [2024-10-29 10:57:54.136193] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:08:48.758 request: 00:08:48.758 { 00:08:48.758 "name": "raid_bdev1", 00:08:48.758 "raid_level": "raid0", 00:08:48.758 "base_bdevs": [ 00:08:48.758 "malloc1", 00:08:48.758 "malloc2", 00:08:48.758 "malloc3" 00:08:48.758 ], 00:08:48.758 "strip_size_kb": 64, 00:08:48.758 "superblock": false, 00:08:48.758 "method": "bdev_raid_create", 00:08:48.758 "req_id": 1 00:08:48.758 } 00:08:48.758 Got JSON-RPC error response 00:08:48.758 response: 00:08:48.758 { 00:08:48.758 "code": -17, 00:08:48.758 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:48.758 } 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 [2024-10-29 10:57:54.201136] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:48.758 [2024-10-29 10:57:54.201261] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:48.758 [2024-10-29 10:57:54.201297] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:48.758 [2024-10-29 10:57:54.201328] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:48.758 [2024-10-29 10:57:54.203880] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:48.758 [2024-10-29 10:57:54.203960] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:48.758 [2024-10-29 10:57:54.204068] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:48.758 [2024-10-29 10:57:54.204133] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:48.758 pt1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.758 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.018 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.018 "name": "raid_bdev1", 00:08:49.018 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:49.018 "strip_size_kb": 64, 00:08:49.018 "state": "configuring", 00:08:49.018 "raid_level": "raid0", 00:08:49.018 "superblock": true, 00:08:49.018 "num_base_bdevs": 3, 00:08:49.018 "num_base_bdevs_discovered": 1, 00:08:49.018 "num_base_bdevs_operational": 3, 00:08:49.018 "base_bdevs_list": [ 00:08:49.018 { 00:08:49.018 "name": "pt1", 00:08:49.018 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:49.018 "is_configured": true, 00:08:49.018 "data_offset": 2048, 00:08:49.018 "data_size": 63488 00:08:49.018 }, 00:08:49.018 { 00:08:49.018 "name": null, 00:08:49.018 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:49.018 "is_configured": false, 00:08:49.019 "data_offset": 2048, 00:08:49.019 "data_size": 63488 00:08:49.019 }, 00:08:49.019 { 00:08:49.019 "name": null, 00:08:49.019 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:49.019 "is_configured": false, 00:08:49.019 "data_offset": 2048, 00:08:49.019 "data_size": 63488 00:08:49.019 } 00:08:49.019 ] 00:08:49.019 }' 00:08:49.019 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.019 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.279 [2024-10-29 10:57:54.608528] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:49.279 [2024-10-29 10:57:54.608678] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.279 [2024-10-29 10:57:54.608716] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:08:49.279 [2024-10-29 10:57:54.608748] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.279 [2024-10-29 10:57:54.609254] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.279 [2024-10-29 10:57:54.609321] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:49.279 [2024-10-29 10:57:54.609461] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:49.279 [2024-10-29 10:57:54.609521] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:49.279 pt2 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.279 [2024-10-29 10:57:54.616497] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.279 "name": "raid_bdev1", 00:08:49.279 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:49.279 "strip_size_kb": 64, 00:08:49.279 "state": "configuring", 00:08:49.279 "raid_level": "raid0", 00:08:49.279 "superblock": true, 00:08:49.279 "num_base_bdevs": 3, 00:08:49.279 "num_base_bdevs_discovered": 1, 00:08:49.279 "num_base_bdevs_operational": 3, 00:08:49.279 "base_bdevs_list": [ 00:08:49.279 { 00:08:49.279 "name": "pt1", 00:08:49.279 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:49.279 "is_configured": true, 00:08:49.279 "data_offset": 2048, 00:08:49.279 "data_size": 63488 00:08:49.279 }, 00:08:49.279 { 00:08:49.279 "name": null, 00:08:49.279 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:49.279 "is_configured": false, 00:08:49.279 "data_offset": 0, 00:08:49.279 "data_size": 63488 00:08:49.279 }, 00:08:49.279 { 00:08:49.279 "name": null, 00:08:49.279 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:49.279 "is_configured": false, 00:08:49.279 "data_offset": 2048, 00:08:49.279 "data_size": 63488 00:08:49.279 } 00:08:49.279 ] 00:08:49.279 }' 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.279 10:57:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.850 [2024-10-29 10:57:55.063793] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:49.850 [2024-10-29 10:57:55.063896] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.850 [2024-10-29 10:57:55.063923] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:08:49.850 [2024-10-29 10:57:55.063934] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.850 [2024-10-29 10:57:55.064443] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.850 [2024-10-29 10:57:55.064464] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:49.850 [2024-10-29 10:57:55.064575] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:49.850 [2024-10-29 10:57:55.064612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:49.850 pt2 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.850 [2024-10-29 10:57:55.075720] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:49.850 [2024-10-29 10:57:55.075838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.850 [2024-10-29 10:57:55.075866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:08:49.850 [2024-10-29 10:57:55.075875] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.850 [2024-10-29 10:57:55.076322] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.850 [2024-10-29 10:57:55.076339] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:49.850 [2024-10-29 10:57:55.076433] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:49.850 [2024-10-29 10:57:55.076454] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:49.850 [2024-10-29 10:57:55.076564] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:08:49.850 [2024-10-29 10:57:55.076572] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:49.850 [2024-10-29 10:57:55.076857] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:08:49.850 [2024-10-29 10:57:55.076987] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:08:49.850 [2024-10-29 10:57:55.076999] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:08:49.850 [2024-10-29 10:57:55.077112] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.850 pt3 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.850 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.851 "name": "raid_bdev1", 00:08:49.851 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:49.851 "strip_size_kb": 64, 00:08:49.851 "state": "online", 00:08:49.851 "raid_level": "raid0", 00:08:49.851 "superblock": true, 00:08:49.851 "num_base_bdevs": 3, 00:08:49.851 "num_base_bdevs_discovered": 3, 00:08:49.851 "num_base_bdevs_operational": 3, 00:08:49.851 "base_bdevs_list": [ 00:08:49.851 { 00:08:49.851 "name": "pt1", 00:08:49.851 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:49.851 "is_configured": true, 00:08:49.851 "data_offset": 2048, 00:08:49.851 "data_size": 63488 00:08:49.851 }, 00:08:49.851 { 00:08:49.851 "name": "pt2", 00:08:49.851 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:49.851 "is_configured": true, 00:08:49.851 "data_offset": 2048, 00:08:49.851 "data_size": 63488 00:08:49.851 }, 00:08:49.851 { 00:08:49.851 "name": "pt3", 00:08:49.851 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:49.851 "is_configured": true, 00:08:49.851 "data_offset": 2048, 00:08:49.851 "data_size": 63488 00:08:49.851 } 00:08:49.851 ] 00:08:49.851 }' 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.851 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:50.111 [2024-10-29 10:57:55.499455] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.111 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:50.111 "name": "raid_bdev1", 00:08:50.111 "aliases": [ 00:08:50.111 "d6612ea0-5405-4cff-80c0-511cd015af02" 00:08:50.111 ], 00:08:50.111 "product_name": "Raid Volume", 00:08:50.111 "block_size": 512, 00:08:50.111 "num_blocks": 190464, 00:08:50.111 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:50.111 "assigned_rate_limits": { 00:08:50.111 "rw_ios_per_sec": 0, 00:08:50.111 "rw_mbytes_per_sec": 0, 00:08:50.111 "r_mbytes_per_sec": 0, 00:08:50.111 "w_mbytes_per_sec": 0 00:08:50.111 }, 00:08:50.111 "claimed": false, 00:08:50.111 "zoned": false, 00:08:50.111 "supported_io_types": { 00:08:50.111 "read": true, 00:08:50.111 "write": true, 00:08:50.111 "unmap": true, 00:08:50.111 "flush": true, 00:08:50.111 "reset": true, 00:08:50.111 "nvme_admin": false, 00:08:50.111 "nvme_io": false, 00:08:50.111 "nvme_io_md": false, 00:08:50.111 "write_zeroes": true, 00:08:50.111 "zcopy": false, 00:08:50.111 "get_zone_info": false, 00:08:50.111 "zone_management": false, 00:08:50.111 "zone_append": false, 00:08:50.111 "compare": false, 00:08:50.111 "compare_and_write": false, 00:08:50.111 "abort": false, 00:08:50.111 "seek_hole": false, 00:08:50.111 "seek_data": false, 00:08:50.111 "copy": false, 00:08:50.111 "nvme_iov_md": false 00:08:50.111 }, 00:08:50.111 "memory_domains": [ 00:08:50.111 { 00:08:50.111 "dma_device_id": "system", 00:08:50.111 "dma_device_type": 1 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.111 "dma_device_type": 2 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "dma_device_id": "system", 00:08:50.111 "dma_device_type": 1 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.111 "dma_device_type": 2 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "dma_device_id": "system", 00:08:50.111 "dma_device_type": 1 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.111 "dma_device_type": 2 00:08:50.111 } 00:08:50.111 ], 00:08:50.111 "driver_specific": { 00:08:50.111 "raid": { 00:08:50.111 "uuid": "d6612ea0-5405-4cff-80c0-511cd015af02", 00:08:50.111 "strip_size_kb": 64, 00:08:50.111 "state": "online", 00:08:50.111 "raid_level": "raid0", 00:08:50.111 "superblock": true, 00:08:50.111 "num_base_bdevs": 3, 00:08:50.111 "num_base_bdevs_discovered": 3, 00:08:50.111 "num_base_bdevs_operational": 3, 00:08:50.111 "base_bdevs_list": [ 00:08:50.111 { 00:08:50.111 "name": "pt1", 00:08:50.111 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:50.111 "is_configured": true, 00:08:50.111 "data_offset": 2048, 00:08:50.111 "data_size": 63488 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "name": "pt2", 00:08:50.111 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:50.111 "is_configured": true, 00:08:50.111 "data_offset": 2048, 00:08:50.111 "data_size": 63488 00:08:50.111 }, 00:08:50.111 { 00:08:50.111 "name": "pt3", 00:08:50.111 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:50.111 "is_configured": true, 00:08:50.111 "data_offset": 2048, 00:08:50.111 "data_size": 63488 00:08:50.111 } 00:08:50.111 ] 00:08:50.111 } 00:08:50.111 } 00:08:50.111 }' 00:08:50.112 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:50.112 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:50.112 pt2 00:08:50.112 pt3' 00:08:50.112 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.372 [2024-10-29 10:57:55.782783] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' d6612ea0-5405-4cff-80c0-511cd015af02 '!=' d6612ea0-5405-4cff-80c0-511cd015af02 ']' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 76384 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 76384 ']' 00:08:50.372 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 76384 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 76384 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 76384' 00:08:50.373 killing process with pid 76384 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 76384 00:08:50.373 [2024-10-29 10:57:55.851904] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:50.373 10:57:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 76384 00:08:50.373 [2024-10-29 10:57:55.852094] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:50.373 [2024-10-29 10:57:55.852175] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:50.373 [2024-10-29 10:57:55.852246] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:08:50.632 [2024-10-29 10:57:55.913343] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:50.950 10:57:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:50.950 00:08:50.950 real 0m3.961s 00:08:50.950 user 0m6.042s 00:08:50.950 sys 0m0.899s 00:08:50.950 10:57:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:50.950 10:57:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.950 ************************************ 00:08:50.950 END TEST raid_superblock_test 00:08:50.950 ************************************ 00:08:50.950 10:57:56 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:50.950 10:57:56 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:50.950 10:57:56 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:50.950 10:57:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:50.950 ************************************ 00:08:50.950 START TEST raid_read_error_test 00:08:50.950 ************************************ 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 3 read 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.xwZ9j4UZli 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76625 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76625 00:08:50.950 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 76625 ']' 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:50.950 10:57:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.950 [2024-10-29 10:57:56.415301] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:50.950 [2024-10-29 10:57:56.415433] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76625 ] 00:08:51.229 [2024-10-29 10:57:56.563612] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:51.229 [2024-10-29 10:57:56.601589] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:51.229 [2024-10-29 10:57:56.678230] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:51.229 [2024-10-29 10:57:56.678364] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.799 BaseBdev1_malloc 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.799 true 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.799 [2024-10-29 10:57:57.292670] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:51.799 [2024-10-29 10:57:57.292781] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:51.799 [2024-10-29 10:57:57.292826] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:51.799 [2024-10-29 10:57:57.292879] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:51.799 [2024-10-29 10:57:57.295422] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:51.799 [2024-10-29 10:57:57.295489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:51.799 BaseBdev1 00:08:51.799 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 BaseBdev2_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 true 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 [2024-10-29 10:57:57.339481] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:52.060 [2024-10-29 10:57:57.339529] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.060 [2024-10-29 10:57:57.339547] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:52.060 [2024-10-29 10:57:57.339556] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.060 [2024-10-29 10:57:57.341836] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.060 [2024-10-29 10:57:57.341919] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:52.060 BaseBdev2 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 BaseBdev3_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 true 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 [2024-10-29 10:57:57.386077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:52.060 [2024-10-29 10:57:57.386123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.060 [2024-10-29 10:57:57.386142] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:52.060 [2024-10-29 10:57:57.386152] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.060 [2024-10-29 10:57:57.388504] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.060 [2024-10-29 10:57:57.388539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:52.060 BaseBdev3 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 [2024-10-29 10:57:57.398116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.060 [2024-10-29 10:57:57.400183] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:52.060 [2024-10-29 10:57:57.400327] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:52.060 [2024-10-29 10:57:57.400533] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:52.060 [2024-10-29 10:57:57.400560] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:52.060 [2024-10-29 10:57:57.400788] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:52.060 [2024-10-29 10:57:57.400943] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:52.060 [2024-10-29 10:57:57.400953] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:52.060 [2024-10-29 10:57:57.401072] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.060 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.060 "name": "raid_bdev1", 00:08:52.060 "uuid": "c103cdc9-6f92-47e7-b85b-b6f52fd3e4e1", 00:08:52.060 "strip_size_kb": 64, 00:08:52.060 "state": "online", 00:08:52.060 "raid_level": "raid0", 00:08:52.060 "superblock": true, 00:08:52.060 "num_base_bdevs": 3, 00:08:52.060 "num_base_bdevs_discovered": 3, 00:08:52.060 "num_base_bdevs_operational": 3, 00:08:52.060 "base_bdevs_list": [ 00:08:52.060 { 00:08:52.060 "name": "BaseBdev1", 00:08:52.060 "uuid": "e1227f83-19c6-535e-ab5d-09cf1882de43", 00:08:52.060 "is_configured": true, 00:08:52.060 "data_offset": 2048, 00:08:52.060 "data_size": 63488 00:08:52.060 }, 00:08:52.060 { 00:08:52.060 "name": "BaseBdev2", 00:08:52.060 "uuid": "bb124de3-9ba1-5159-9ee7-689dcd2b63b4", 00:08:52.060 "is_configured": true, 00:08:52.061 "data_offset": 2048, 00:08:52.061 "data_size": 63488 00:08:52.061 }, 00:08:52.061 { 00:08:52.061 "name": "BaseBdev3", 00:08:52.061 "uuid": "130d4402-c22a-5527-a69a-8a7b4b3d64f2", 00:08:52.061 "is_configured": true, 00:08:52.061 "data_offset": 2048, 00:08:52.061 "data_size": 63488 00:08:52.061 } 00:08:52.061 ] 00:08:52.061 }' 00:08:52.061 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.061 10:57:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.320 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:52.320 10:57:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:52.580 [2024-10-29 10:57:57.901660] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.520 "name": "raid_bdev1", 00:08:53.520 "uuid": "c103cdc9-6f92-47e7-b85b-b6f52fd3e4e1", 00:08:53.520 "strip_size_kb": 64, 00:08:53.520 "state": "online", 00:08:53.520 "raid_level": "raid0", 00:08:53.520 "superblock": true, 00:08:53.520 "num_base_bdevs": 3, 00:08:53.520 "num_base_bdevs_discovered": 3, 00:08:53.520 "num_base_bdevs_operational": 3, 00:08:53.520 "base_bdevs_list": [ 00:08:53.520 { 00:08:53.520 "name": "BaseBdev1", 00:08:53.520 "uuid": "e1227f83-19c6-535e-ab5d-09cf1882de43", 00:08:53.520 "is_configured": true, 00:08:53.520 "data_offset": 2048, 00:08:53.520 "data_size": 63488 00:08:53.520 }, 00:08:53.520 { 00:08:53.520 "name": "BaseBdev2", 00:08:53.520 "uuid": "bb124de3-9ba1-5159-9ee7-689dcd2b63b4", 00:08:53.520 "is_configured": true, 00:08:53.520 "data_offset": 2048, 00:08:53.520 "data_size": 63488 00:08:53.520 }, 00:08:53.520 { 00:08:53.520 "name": "BaseBdev3", 00:08:53.520 "uuid": "130d4402-c22a-5527-a69a-8a7b4b3d64f2", 00:08:53.520 "is_configured": true, 00:08:53.520 "data_offset": 2048, 00:08:53.520 "data_size": 63488 00:08:53.520 } 00:08:53.520 ] 00:08:53.520 }' 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.520 10:57:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.781 [2024-10-29 10:57:59.246639] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:53.781 [2024-10-29 10:57:59.246690] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:53.781 [2024-10-29 10:57:59.249231] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:53.781 [2024-10-29 10:57:59.249300] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:53.781 [2024-10-29 10:57:59.249343] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:53.781 [2024-10-29 10:57:59.249356] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:53.781 { 00:08:53.781 "results": [ 00:08:53.781 { 00:08:53.781 "job": "raid_bdev1", 00:08:53.781 "core_mask": "0x1", 00:08:53.781 "workload": "randrw", 00:08:53.781 "percentage": 50, 00:08:53.781 "status": "finished", 00:08:53.781 "queue_depth": 1, 00:08:53.781 "io_size": 131072, 00:08:53.781 "runtime": 1.345443, 00:08:53.781 "iops": 14735.666988493753, 00:08:53.781 "mibps": 1841.958373561719, 00:08:53.781 "io_failed": 1, 00:08:53.781 "io_timeout": 0, 00:08:53.781 "avg_latency_us": 95.24922915093286, 00:08:53.781 "min_latency_us": 25.6, 00:08:53.781 "max_latency_us": 1531.0812227074236 00:08:53.781 } 00:08:53.781 ], 00:08:53.781 "core_count": 1 00:08:53.781 } 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76625 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 76625 ']' 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 76625 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:53.781 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 76625 00:08:54.042 killing process with pid 76625 00:08:54.042 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:54.042 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:54.042 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 76625' 00:08:54.042 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 76625 00:08:54.042 [2024-10-29 10:57:59.294424] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:54.042 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 76625 00:08:54.042 [2024-10-29 10:57:59.340106] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.xwZ9j4UZli 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:54.301 ************************************ 00:08:54.301 END TEST raid_read_error_test 00:08:54.301 ************************************ 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:54.301 00:08:54.301 real 0m3.352s 00:08:54.301 user 0m4.107s 00:08:54.301 sys 0m0.602s 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:54.301 10:57:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.301 10:57:59 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:54.301 10:57:59 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:54.301 10:57:59 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:54.301 10:57:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:54.301 ************************************ 00:08:54.301 START TEST raid_write_error_test 00:08:54.301 ************************************ 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 3 write 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ymvEVicOz4 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76755 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76755 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 76755 ']' 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:54.301 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:54.301 10:57:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.560 [2024-10-29 10:57:59.830582] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:54.560 [2024-10-29 10:57:59.830713] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76755 ] 00:08:54.561 [2024-10-29 10:58:00.001719] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:54.561 [2024-10-29 10:58:00.045659] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:54.820 [2024-10-29 10:58:00.122554] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:54.820 [2024-10-29 10:58:00.122598] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 BaseBdev1_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 true 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 [2024-10-29 10:58:00.741082] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:55.393 [2024-10-29 10:58:00.741239] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:55.393 [2024-10-29 10:58:00.741287] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:08:55.393 [2024-10-29 10:58:00.741317] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:55.393 [2024-10-29 10:58:00.743839] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:55.393 [2024-10-29 10:58:00.743920] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:55.393 BaseBdev1 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 BaseBdev2_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 true 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 [2024-10-29 10:58:00.779881] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:55.393 [2024-10-29 10:58:00.780004] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:55.393 [2024-10-29 10:58:00.780027] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:55.393 [2024-10-29 10:58:00.780038] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:55.393 [2024-10-29 10:58:00.782421] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:55.393 [2024-10-29 10:58:00.782455] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:55.393 BaseBdev2 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 BaseBdev3_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 true 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 [2024-10-29 10:58:00.814586] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:55.393 [2024-10-29 10:58:00.814698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:55.393 [2024-10-29 10:58:00.814723] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:55.393 [2024-10-29 10:58:00.814733] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:55.393 [2024-10-29 10:58:00.817134] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:55.393 [2024-10-29 10:58:00.817169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:55.393 BaseBdev3 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.393 [2024-10-29 10:58:00.822649] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:55.393 [2024-10-29 10:58:00.824781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:55.393 [2024-10-29 10:58:00.824860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:55.393 [2024-10-29 10:58:00.825035] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:08:55.393 [2024-10-29 10:58:00.825049] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:55.393 [2024-10-29 10:58:00.825276] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:08:55.393 [2024-10-29 10:58:00.825427] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:08:55.393 [2024-10-29 10:58:00.825438] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:08:55.393 [2024-10-29 10:58:00.825591] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:55.393 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.394 "name": "raid_bdev1", 00:08:55.394 "uuid": "fc5d8446-a3de-44bb-8aa8-3c0780f92805", 00:08:55.394 "strip_size_kb": 64, 00:08:55.394 "state": "online", 00:08:55.394 "raid_level": "raid0", 00:08:55.394 "superblock": true, 00:08:55.394 "num_base_bdevs": 3, 00:08:55.394 "num_base_bdevs_discovered": 3, 00:08:55.394 "num_base_bdevs_operational": 3, 00:08:55.394 "base_bdevs_list": [ 00:08:55.394 { 00:08:55.394 "name": "BaseBdev1", 00:08:55.394 "uuid": "5bafd0c7-1fbf-594c-868d-5677e875119d", 00:08:55.394 "is_configured": true, 00:08:55.394 "data_offset": 2048, 00:08:55.394 "data_size": 63488 00:08:55.394 }, 00:08:55.394 { 00:08:55.394 "name": "BaseBdev2", 00:08:55.394 "uuid": "93f3fb71-5ee6-56ab-a701-1c21600b055e", 00:08:55.394 "is_configured": true, 00:08:55.394 "data_offset": 2048, 00:08:55.394 "data_size": 63488 00:08:55.394 }, 00:08:55.394 { 00:08:55.394 "name": "BaseBdev3", 00:08:55.394 "uuid": "6df3b7a1-7501-5703-aafe-e600b9316fb1", 00:08:55.394 "is_configured": true, 00:08:55.394 "data_offset": 2048, 00:08:55.394 "data_size": 63488 00:08:55.394 } 00:08:55.394 ] 00:08:55.394 }' 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.394 10:58:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.963 10:58:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:55.963 10:58:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:55.963 [2024-10-29 10:58:01.378200] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.901 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.902 "name": "raid_bdev1", 00:08:56.902 "uuid": "fc5d8446-a3de-44bb-8aa8-3c0780f92805", 00:08:56.902 "strip_size_kb": 64, 00:08:56.902 "state": "online", 00:08:56.902 "raid_level": "raid0", 00:08:56.902 "superblock": true, 00:08:56.902 "num_base_bdevs": 3, 00:08:56.902 "num_base_bdevs_discovered": 3, 00:08:56.902 "num_base_bdevs_operational": 3, 00:08:56.902 "base_bdevs_list": [ 00:08:56.902 { 00:08:56.902 "name": "BaseBdev1", 00:08:56.902 "uuid": "5bafd0c7-1fbf-594c-868d-5677e875119d", 00:08:56.902 "is_configured": true, 00:08:56.902 "data_offset": 2048, 00:08:56.902 "data_size": 63488 00:08:56.902 }, 00:08:56.902 { 00:08:56.902 "name": "BaseBdev2", 00:08:56.902 "uuid": "93f3fb71-5ee6-56ab-a701-1c21600b055e", 00:08:56.902 "is_configured": true, 00:08:56.902 "data_offset": 2048, 00:08:56.902 "data_size": 63488 00:08:56.902 }, 00:08:56.902 { 00:08:56.902 "name": "BaseBdev3", 00:08:56.902 "uuid": "6df3b7a1-7501-5703-aafe-e600b9316fb1", 00:08:56.902 "is_configured": true, 00:08:56.902 "data_offset": 2048, 00:08:56.902 "data_size": 63488 00:08:56.902 } 00:08:56.902 ] 00:08:56.902 }' 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.902 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.471 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:57.471 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.471 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.471 [2024-10-29 10:58:02.702632] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:57.471 [2024-10-29 10:58:02.702690] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:57.471 [2024-10-29 10:58:02.705293] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:57.471 [2024-10-29 10:58:02.705449] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.471 [2024-10-29 10:58:02.705499] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:57.471 [2024-10-29 10:58:02.705522] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:08:57.471 { 00:08:57.471 "results": [ 00:08:57.471 { 00:08:57.471 "job": "raid_bdev1", 00:08:57.472 "core_mask": "0x1", 00:08:57.472 "workload": "randrw", 00:08:57.472 "percentage": 50, 00:08:57.472 "status": "finished", 00:08:57.472 "queue_depth": 1, 00:08:57.472 "io_size": 131072, 00:08:57.472 "runtime": 1.32467, 00:08:57.472 "iops": 14086.527210550552, 00:08:57.472 "mibps": 1760.815901318819, 00:08:57.472 "io_failed": 1, 00:08:57.472 "io_timeout": 0, 00:08:57.472 "avg_latency_us": 99.7111418180831, 00:08:57.472 "min_latency_us": 20.90480349344978, 00:08:57.472 "max_latency_us": 1459.5353711790392 00:08:57.472 } 00:08:57.472 ], 00:08:57.472 "core_count": 1 00:08:57.472 } 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76755 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 76755 ']' 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 76755 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 76755 00:08:57.472 killing process with pid 76755 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 76755' 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 76755 00:08:57.472 [2024-10-29 10:58:02.752882] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:57.472 10:58:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 76755 00:08:57.472 [2024-10-29 10:58:02.801396] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ymvEVicOz4 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:57.731 ************************************ 00:08:57.731 END TEST raid_write_error_test 00:08:57.731 ************************************ 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:08:57.731 00:08:57.731 real 0m3.411s 00:08:57.731 user 0m4.239s 00:08:57.731 sys 0m0.587s 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:08:57.731 10:58:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.731 10:58:03 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:57.731 10:58:03 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:57.731 10:58:03 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:08:57.731 10:58:03 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:08:57.731 10:58:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:57.731 ************************************ 00:08:57.731 START TEST raid_state_function_test 00:08:57.731 ************************************ 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 3 false 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:57.731 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76882 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76882' 00:08:57.732 Process raid pid: 76882 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76882 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 76882 ']' 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:57.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:08:57.732 10:58:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.991 [2024-10-29 10:58:03.300471] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:08:57.991 [2024-10-29 10:58:03.301034] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:57.991 [2024-10-29 10:58:03.469689] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:58.252 [2024-10-29 10:58:03.510455] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:58.252 [2024-10-29 10:58:03.587661] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:58.252 [2024-10-29 10:58:03.587797] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.822 [2024-10-29 10:58:04.131019] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:58.822 [2024-10-29 10:58:04.131141] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:58.822 [2024-10-29 10:58:04.131170] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:58.822 [2024-10-29 10:58:04.131202] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:58.822 [2024-10-29 10:58:04.131225] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:58.822 [2024-10-29 10:58:04.131250] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.822 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.822 "name": "Existed_Raid", 00:08:58.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.822 "strip_size_kb": 64, 00:08:58.822 "state": "configuring", 00:08:58.822 "raid_level": "concat", 00:08:58.822 "superblock": false, 00:08:58.822 "num_base_bdevs": 3, 00:08:58.822 "num_base_bdevs_discovered": 0, 00:08:58.822 "num_base_bdevs_operational": 3, 00:08:58.822 "base_bdevs_list": [ 00:08:58.822 { 00:08:58.822 "name": "BaseBdev1", 00:08:58.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.822 "is_configured": false, 00:08:58.822 "data_offset": 0, 00:08:58.822 "data_size": 0 00:08:58.822 }, 00:08:58.822 { 00:08:58.822 "name": "BaseBdev2", 00:08:58.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.822 "is_configured": false, 00:08:58.822 "data_offset": 0, 00:08:58.822 "data_size": 0 00:08:58.822 }, 00:08:58.822 { 00:08:58.822 "name": "BaseBdev3", 00:08:58.823 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.823 "is_configured": false, 00:08:58.823 "data_offset": 0, 00:08:58.823 "data_size": 0 00:08:58.823 } 00:08:58.823 ] 00:08:58.823 }' 00:08:58.823 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.823 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.083 [2024-10-29 10:58:04.570174] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:59.083 [2024-10-29 10:58:04.570223] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.083 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.344 [2024-10-29 10:58:04.582156] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:59.344 [2024-10-29 10:58:04.582204] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:59.344 [2024-10-29 10:58:04.582212] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:59.344 [2024-10-29 10:58:04.582222] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:59.344 [2024-10-29 10:58:04.582228] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:59.344 [2024-10-29 10:58:04.582238] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.344 [2024-10-29 10:58:04.609490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:59.344 BaseBdev1 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.344 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.344 [ 00:08:59.344 { 00:08:59.344 "name": "BaseBdev1", 00:08:59.344 "aliases": [ 00:08:59.344 "07ada808-4deb-4178-94e9-d80178c3f813" 00:08:59.344 ], 00:08:59.344 "product_name": "Malloc disk", 00:08:59.344 "block_size": 512, 00:08:59.344 "num_blocks": 65536, 00:08:59.344 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:08:59.344 "assigned_rate_limits": { 00:08:59.344 "rw_ios_per_sec": 0, 00:08:59.344 "rw_mbytes_per_sec": 0, 00:08:59.344 "r_mbytes_per_sec": 0, 00:08:59.344 "w_mbytes_per_sec": 0 00:08:59.344 }, 00:08:59.344 "claimed": true, 00:08:59.344 "claim_type": "exclusive_write", 00:08:59.344 "zoned": false, 00:08:59.344 "supported_io_types": { 00:08:59.344 "read": true, 00:08:59.344 "write": true, 00:08:59.344 "unmap": true, 00:08:59.344 "flush": true, 00:08:59.344 "reset": true, 00:08:59.344 "nvme_admin": false, 00:08:59.345 "nvme_io": false, 00:08:59.345 "nvme_io_md": false, 00:08:59.345 "write_zeroes": true, 00:08:59.345 "zcopy": true, 00:08:59.345 "get_zone_info": false, 00:08:59.345 "zone_management": false, 00:08:59.345 "zone_append": false, 00:08:59.345 "compare": false, 00:08:59.345 "compare_and_write": false, 00:08:59.345 "abort": true, 00:08:59.345 "seek_hole": false, 00:08:59.345 "seek_data": false, 00:08:59.345 "copy": true, 00:08:59.345 "nvme_iov_md": false 00:08:59.345 }, 00:08:59.345 "memory_domains": [ 00:08:59.345 { 00:08:59.345 "dma_device_id": "system", 00:08:59.345 "dma_device_type": 1 00:08:59.345 }, 00:08:59.345 { 00:08:59.345 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.345 "dma_device_type": 2 00:08:59.345 } 00:08:59.345 ], 00:08:59.345 "driver_specific": {} 00:08:59.345 } 00:08:59.345 ] 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.345 "name": "Existed_Raid", 00:08:59.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.345 "strip_size_kb": 64, 00:08:59.345 "state": "configuring", 00:08:59.345 "raid_level": "concat", 00:08:59.345 "superblock": false, 00:08:59.345 "num_base_bdevs": 3, 00:08:59.345 "num_base_bdevs_discovered": 1, 00:08:59.345 "num_base_bdevs_operational": 3, 00:08:59.345 "base_bdevs_list": [ 00:08:59.345 { 00:08:59.345 "name": "BaseBdev1", 00:08:59.345 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:08:59.345 "is_configured": true, 00:08:59.345 "data_offset": 0, 00:08:59.345 "data_size": 65536 00:08:59.345 }, 00:08:59.345 { 00:08:59.345 "name": "BaseBdev2", 00:08:59.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.345 "is_configured": false, 00:08:59.345 "data_offset": 0, 00:08:59.345 "data_size": 0 00:08:59.345 }, 00:08:59.345 { 00:08:59.345 "name": "BaseBdev3", 00:08:59.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.345 "is_configured": false, 00:08:59.345 "data_offset": 0, 00:08:59.345 "data_size": 0 00:08:59.345 } 00:08:59.345 ] 00:08:59.345 }' 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.345 10:58:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.916 [2024-10-29 10:58:05.140583] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:59.916 [2024-10-29 10:58:05.140693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.916 [2024-10-29 10:58:05.152601] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:59.916 [2024-10-29 10:58:05.154761] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:59.916 [2024-10-29 10:58:05.154848] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:59.916 [2024-10-29 10:58:05.154861] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:59.916 [2024-10-29 10:58:05.154883] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.916 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.917 "name": "Existed_Raid", 00:08:59.917 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.917 "strip_size_kb": 64, 00:08:59.917 "state": "configuring", 00:08:59.917 "raid_level": "concat", 00:08:59.917 "superblock": false, 00:08:59.917 "num_base_bdevs": 3, 00:08:59.917 "num_base_bdevs_discovered": 1, 00:08:59.917 "num_base_bdevs_operational": 3, 00:08:59.917 "base_bdevs_list": [ 00:08:59.917 { 00:08:59.917 "name": "BaseBdev1", 00:08:59.917 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:08:59.917 "is_configured": true, 00:08:59.917 "data_offset": 0, 00:08:59.917 "data_size": 65536 00:08:59.917 }, 00:08:59.917 { 00:08:59.917 "name": "BaseBdev2", 00:08:59.917 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.917 "is_configured": false, 00:08:59.917 "data_offset": 0, 00:08:59.917 "data_size": 0 00:08:59.917 }, 00:08:59.917 { 00:08:59.917 "name": "BaseBdev3", 00:08:59.917 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.917 "is_configured": false, 00:08:59.917 "data_offset": 0, 00:08:59.917 "data_size": 0 00:08:59.917 } 00:08:59.917 ] 00:08:59.917 }' 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.917 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.177 [2024-10-29 10:58:05.512716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:00.177 BaseBdev2 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.177 [ 00:09:00.177 { 00:09:00.177 "name": "BaseBdev2", 00:09:00.177 "aliases": [ 00:09:00.177 "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0" 00:09:00.177 ], 00:09:00.177 "product_name": "Malloc disk", 00:09:00.177 "block_size": 512, 00:09:00.177 "num_blocks": 65536, 00:09:00.177 "uuid": "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0", 00:09:00.177 "assigned_rate_limits": { 00:09:00.177 "rw_ios_per_sec": 0, 00:09:00.177 "rw_mbytes_per_sec": 0, 00:09:00.177 "r_mbytes_per_sec": 0, 00:09:00.177 "w_mbytes_per_sec": 0 00:09:00.177 }, 00:09:00.177 "claimed": true, 00:09:00.177 "claim_type": "exclusive_write", 00:09:00.177 "zoned": false, 00:09:00.177 "supported_io_types": { 00:09:00.177 "read": true, 00:09:00.177 "write": true, 00:09:00.177 "unmap": true, 00:09:00.177 "flush": true, 00:09:00.177 "reset": true, 00:09:00.177 "nvme_admin": false, 00:09:00.177 "nvme_io": false, 00:09:00.177 "nvme_io_md": false, 00:09:00.177 "write_zeroes": true, 00:09:00.177 "zcopy": true, 00:09:00.177 "get_zone_info": false, 00:09:00.177 "zone_management": false, 00:09:00.177 "zone_append": false, 00:09:00.177 "compare": false, 00:09:00.177 "compare_and_write": false, 00:09:00.177 "abort": true, 00:09:00.177 "seek_hole": false, 00:09:00.177 "seek_data": false, 00:09:00.177 "copy": true, 00:09:00.177 "nvme_iov_md": false 00:09:00.177 }, 00:09:00.177 "memory_domains": [ 00:09:00.177 { 00:09:00.177 "dma_device_id": "system", 00:09:00.177 "dma_device_type": 1 00:09:00.177 }, 00:09:00.177 { 00:09:00.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.177 "dma_device_type": 2 00:09:00.177 } 00:09:00.177 ], 00:09:00.177 "driver_specific": {} 00:09:00.177 } 00:09:00.177 ] 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:00.177 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.178 "name": "Existed_Raid", 00:09:00.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.178 "strip_size_kb": 64, 00:09:00.178 "state": "configuring", 00:09:00.178 "raid_level": "concat", 00:09:00.178 "superblock": false, 00:09:00.178 "num_base_bdevs": 3, 00:09:00.178 "num_base_bdevs_discovered": 2, 00:09:00.178 "num_base_bdevs_operational": 3, 00:09:00.178 "base_bdevs_list": [ 00:09:00.178 { 00:09:00.178 "name": "BaseBdev1", 00:09:00.178 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:09:00.178 "is_configured": true, 00:09:00.178 "data_offset": 0, 00:09:00.178 "data_size": 65536 00:09:00.178 }, 00:09:00.178 { 00:09:00.178 "name": "BaseBdev2", 00:09:00.178 "uuid": "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0", 00:09:00.178 "is_configured": true, 00:09:00.178 "data_offset": 0, 00:09:00.178 "data_size": 65536 00:09:00.178 }, 00:09:00.178 { 00:09:00.178 "name": "BaseBdev3", 00:09:00.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.178 "is_configured": false, 00:09:00.178 "data_offset": 0, 00:09:00.178 "data_size": 0 00:09:00.178 } 00:09:00.178 ] 00:09:00.178 }' 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.178 10:58:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.749 10:58:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.749 [2024-10-29 10:58:06.028585] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:00.749 [2024-10-29 10:58:06.028646] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:00.749 [2024-10-29 10:58:06.028663] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:00.749 [2024-10-29 10:58:06.029001] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:00.749 [2024-10-29 10:58:06.029162] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:00.749 [2024-10-29 10:58:06.029173] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:00.749 [2024-10-29 10:58:06.029429] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:00.749 BaseBdev3 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.749 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.749 [ 00:09:00.749 { 00:09:00.749 "name": "BaseBdev3", 00:09:00.749 "aliases": [ 00:09:00.749 "5aea2858-2054-4a8b-8191-07e0bcfefc25" 00:09:00.749 ], 00:09:00.749 "product_name": "Malloc disk", 00:09:00.749 "block_size": 512, 00:09:00.749 "num_blocks": 65536, 00:09:00.749 "uuid": "5aea2858-2054-4a8b-8191-07e0bcfefc25", 00:09:00.749 "assigned_rate_limits": { 00:09:00.749 "rw_ios_per_sec": 0, 00:09:00.750 "rw_mbytes_per_sec": 0, 00:09:00.750 "r_mbytes_per_sec": 0, 00:09:00.750 "w_mbytes_per_sec": 0 00:09:00.750 }, 00:09:00.750 "claimed": true, 00:09:00.750 "claim_type": "exclusive_write", 00:09:00.750 "zoned": false, 00:09:00.750 "supported_io_types": { 00:09:00.750 "read": true, 00:09:00.750 "write": true, 00:09:00.750 "unmap": true, 00:09:00.750 "flush": true, 00:09:00.750 "reset": true, 00:09:00.750 "nvme_admin": false, 00:09:00.750 "nvme_io": false, 00:09:00.750 "nvme_io_md": false, 00:09:00.750 "write_zeroes": true, 00:09:00.750 "zcopy": true, 00:09:00.750 "get_zone_info": false, 00:09:00.750 "zone_management": false, 00:09:00.750 "zone_append": false, 00:09:00.750 "compare": false, 00:09:00.750 "compare_and_write": false, 00:09:00.750 "abort": true, 00:09:00.750 "seek_hole": false, 00:09:00.750 "seek_data": false, 00:09:00.750 "copy": true, 00:09:00.750 "nvme_iov_md": false 00:09:00.750 }, 00:09:00.750 "memory_domains": [ 00:09:00.750 { 00:09:00.750 "dma_device_id": "system", 00:09:00.750 "dma_device_type": 1 00:09:00.750 }, 00:09:00.750 { 00:09:00.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.750 "dma_device_type": 2 00:09:00.750 } 00:09:00.750 ], 00:09:00.750 "driver_specific": {} 00:09:00.750 } 00:09:00.750 ] 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.750 "name": "Existed_Raid", 00:09:00.750 "uuid": "0992bbde-6b38-420c-9842-6c8ee48bd05d", 00:09:00.750 "strip_size_kb": 64, 00:09:00.750 "state": "online", 00:09:00.750 "raid_level": "concat", 00:09:00.750 "superblock": false, 00:09:00.750 "num_base_bdevs": 3, 00:09:00.750 "num_base_bdevs_discovered": 3, 00:09:00.750 "num_base_bdevs_operational": 3, 00:09:00.750 "base_bdevs_list": [ 00:09:00.750 { 00:09:00.750 "name": "BaseBdev1", 00:09:00.750 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:09:00.750 "is_configured": true, 00:09:00.750 "data_offset": 0, 00:09:00.750 "data_size": 65536 00:09:00.750 }, 00:09:00.750 { 00:09:00.750 "name": "BaseBdev2", 00:09:00.750 "uuid": "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0", 00:09:00.750 "is_configured": true, 00:09:00.750 "data_offset": 0, 00:09:00.750 "data_size": 65536 00:09:00.750 }, 00:09:00.750 { 00:09:00.750 "name": "BaseBdev3", 00:09:00.750 "uuid": "5aea2858-2054-4a8b-8191-07e0bcfefc25", 00:09:00.750 "is_configured": true, 00:09:00.750 "data_offset": 0, 00:09:00.750 "data_size": 65536 00:09:00.750 } 00:09:00.750 ] 00:09:00.750 }' 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.750 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.011 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.011 [2024-10-29 10:58:06.500430] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:01.270 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.270 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:01.270 "name": "Existed_Raid", 00:09:01.270 "aliases": [ 00:09:01.270 "0992bbde-6b38-420c-9842-6c8ee48bd05d" 00:09:01.270 ], 00:09:01.270 "product_name": "Raid Volume", 00:09:01.270 "block_size": 512, 00:09:01.270 "num_blocks": 196608, 00:09:01.270 "uuid": "0992bbde-6b38-420c-9842-6c8ee48bd05d", 00:09:01.270 "assigned_rate_limits": { 00:09:01.270 "rw_ios_per_sec": 0, 00:09:01.270 "rw_mbytes_per_sec": 0, 00:09:01.270 "r_mbytes_per_sec": 0, 00:09:01.270 "w_mbytes_per_sec": 0 00:09:01.270 }, 00:09:01.270 "claimed": false, 00:09:01.270 "zoned": false, 00:09:01.270 "supported_io_types": { 00:09:01.270 "read": true, 00:09:01.270 "write": true, 00:09:01.270 "unmap": true, 00:09:01.270 "flush": true, 00:09:01.270 "reset": true, 00:09:01.270 "nvme_admin": false, 00:09:01.270 "nvme_io": false, 00:09:01.270 "nvme_io_md": false, 00:09:01.270 "write_zeroes": true, 00:09:01.270 "zcopy": false, 00:09:01.270 "get_zone_info": false, 00:09:01.270 "zone_management": false, 00:09:01.270 "zone_append": false, 00:09:01.270 "compare": false, 00:09:01.270 "compare_and_write": false, 00:09:01.270 "abort": false, 00:09:01.270 "seek_hole": false, 00:09:01.270 "seek_data": false, 00:09:01.270 "copy": false, 00:09:01.270 "nvme_iov_md": false 00:09:01.270 }, 00:09:01.270 "memory_domains": [ 00:09:01.270 { 00:09:01.270 "dma_device_id": "system", 00:09:01.270 "dma_device_type": 1 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.270 "dma_device_type": 2 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "dma_device_id": "system", 00:09:01.270 "dma_device_type": 1 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.270 "dma_device_type": 2 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "dma_device_id": "system", 00:09:01.270 "dma_device_type": 1 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.270 "dma_device_type": 2 00:09:01.270 } 00:09:01.270 ], 00:09:01.270 "driver_specific": { 00:09:01.270 "raid": { 00:09:01.270 "uuid": "0992bbde-6b38-420c-9842-6c8ee48bd05d", 00:09:01.270 "strip_size_kb": 64, 00:09:01.270 "state": "online", 00:09:01.270 "raid_level": "concat", 00:09:01.270 "superblock": false, 00:09:01.270 "num_base_bdevs": 3, 00:09:01.270 "num_base_bdevs_discovered": 3, 00:09:01.270 "num_base_bdevs_operational": 3, 00:09:01.270 "base_bdevs_list": [ 00:09:01.270 { 00:09:01.270 "name": "BaseBdev1", 00:09:01.270 "uuid": "07ada808-4deb-4178-94e9-d80178c3f813", 00:09:01.270 "is_configured": true, 00:09:01.270 "data_offset": 0, 00:09:01.270 "data_size": 65536 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "name": "BaseBdev2", 00:09:01.270 "uuid": "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0", 00:09:01.270 "is_configured": true, 00:09:01.270 "data_offset": 0, 00:09:01.270 "data_size": 65536 00:09:01.270 }, 00:09:01.270 { 00:09:01.270 "name": "BaseBdev3", 00:09:01.270 "uuid": "5aea2858-2054-4a8b-8191-07e0bcfefc25", 00:09:01.270 "is_configured": true, 00:09:01.270 "data_offset": 0, 00:09:01.270 "data_size": 65536 00:09:01.270 } 00:09:01.270 ] 00:09:01.270 } 00:09:01.270 } 00:09:01.271 }' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:01.271 BaseBdev2 00:09:01.271 BaseBdev3' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.271 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.531 [2024-10-29 10:58:06.779628] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:01.531 [2024-10-29 10:58:06.779749] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:01.531 [2024-10-29 10:58:06.779843] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.531 "name": "Existed_Raid", 00:09:01.531 "uuid": "0992bbde-6b38-420c-9842-6c8ee48bd05d", 00:09:01.531 "strip_size_kb": 64, 00:09:01.531 "state": "offline", 00:09:01.531 "raid_level": "concat", 00:09:01.531 "superblock": false, 00:09:01.531 "num_base_bdevs": 3, 00:09:01.531 "num_base_bdevs_discovered": 2, 00:09:01.531 "num_base_bdevs_operational": 2, 00:09:01.531 "base_bdevs_list": [ 00:09:01.531 { 00:09:01.531 "name": null, 00:09:01.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.531 "is_configured": false, 00:09:01.531 "data_offset": 0, 00:09:01.531 "data_size": 65536 00:09:01.531 }, 00:09:01.531 { 00:09:01.531 "name": "BaseBdev2", 00:09:01.531 "uuid": "3df97c48-5d8d-4eec-ac4a-79d8cd436bd0", 00:09:01.531 "is_configured": true, 00:09:01.531 "data_offset": 0, 00:09:01.531 "data_size": 65536 00:09:01.531 }, 00:09:01.531 { 00:09:01.531 "name": "BaseBdev3", 00:09:01.531 "uuid": "5aea2858-2054-4a8b-8191-07e0bcfefc25", 00:09:01.531 "is_configured": true, 00:09:01.531 "data_offset": 0, 00:09:01.531 "data_size": 65536 00:09:01.531 } 00:09:01.531 ] 00:09:01.531 }' 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.531 10:58:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.791 [2024-10-29 10:58:07.263529] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:01.791 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 [2024-10-29 10:58:07.340031] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:02.052 [2024-10-29 10:58:07.340175] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 BaseBdev2 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.052 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.052 [ 00:09:02.052 { 00:09:02.052 "name": "BaseBdev2", 00:09:02.053 "aliases": [ 00:09:02.053 "26c1c4ff-004c-420c-8833-a94bd8240128" 00:09:02.053 ], 00:09:02.053 "product_name": "Malloc disk", 00:09:02.053 "block_size": 512, 00:09:02.053 "num_blocks": 65536, 00:09:02.053 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:02.053 "assigned_rate_limits": { 00:09:02.053 "rw_ios_per_sec": 0, 00:09:02.053 "rw_mbytes_per_sec": 0, 00:09:02.053 "r_mbytes_per_sec": 0, 00:09:02.053 "w_mbytes_per_sec": 0 00:09:02.053 }, 00:09:02.053 "claimed": false, 00:09:02.053 "zoned": false, 00:09:02.053 "supported_io_types": { 00:09:02.053 "read": true, 00:09:02.053 "write": true, 00:09:02.053 "unmap": true, 00:09:02.053 "flush": true, 00:09:02.053 "reset": true, 00:09:02.053 "nvme_admin": false, 00:09:02.053 "nvme_io": false, 00:09:02.053 "nvme_io_md": false, 00:09:02.053 "write_zeroes": true, 00:09:02.053 "zcopy": true, 00:09:02.053 "get_zone_info": false, 00:09:02.053 "zone_management": false, 00:09:02.053 "zone_append": false, 00:09:02.053 "compare": false, 00:09:02.053 "compare_and_write": false, 00:09:02.053 "abort": true, 00:09:02.053 "seek_hole": false, 00:09:02.053 "seek_data": false, 00:09:02.053 "copy": true, 00:09:02.053 "nvme_iov_md": false 00:09:02.053 }, 00:09:02.053 "memory_domains": [ 00:09:02.053 { 00:09:02.053 "dma_device_id": "system", 00:09:02.053 "dma_device_type": 1 00:09:02.053 }, 00:09:02.053 { 00:09:02.053 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.053 "dma_device_type": 2 00:09:02.053 } 00:09:02.053 ], 00:09:02.053 "driver_specific": {} 00:09:02.053 } 00:09:02.053 ] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.053 BaseBdev3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.053 [ 00:09:02.053 { 00:09:02.053 "name": "BaseBdev3", 00:09:02.053 "aliases": [ 00:09:02.053 "ef65c798-9cab-4425-ae0b-57b08b1616a9" 00:09:02.053 ], 00:09:02.053 "product_name": "Malloc disk", 00:09:02.053 "block_size": 512, 00:09:02.053 "num_blocks": 65536, 00:09:02.053 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:02.053 "assigned_rate_limits": { 00:09:02.053 "rw_ios_per_sec": 0, 00:09:02.053 "rw_mbytes_per_sec": 0, 00:09:02.053 "r_mbytes_per_sec": 0, 00:09:02.053 "w_mbytes_per_sec": 0 00:09:02.053 }, 00:09:02.053 "claimed": false, 00:09:02.053 "zoned": false, 00:09:02.053 "supported_io_types": { 00:09:02.053 "read": true, 00:09:02.053 "write": true, 00:09:02.053 "unmap": true, 00:09:02.053 "flush": true, 00:09:02.053 "reset": true, 00:09:02.053 "nvme_admin": false, 00:09:02.053 "nvme_io": false, 00:09:02.053 "nvme_io_md": false, 00:09:02.053 "write_zeroes": true, 00:09:02.053 "zcopy": true, 00:09:02.053 "get_zone_info": false, 00:09:02.053 "zone_management": false, 00:09:02.053 "zone_append": false, 00:09:02.053 "compare": false, 00:09:02.053 "compare_and_write": false, 00:09:02.053 "abort": true, 00:09:02.053 "seek_hole": false, 00:09:02.053 "seek_data": false, 00:09:02.053 "copy": true, 00:09:02.053 "nvme_iov_md": false 00:09:02.053 }, 00:09:02.053 "memory_domains": [ 00:09:02.053 { 00:09:02.053 "dma_device_id": "system", 00:09:02.053 "dma_device_type": 1 00:09:02.053 }, 00:09:02.053 { 00:09:02.053 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.053 "dma_device_type": 2 00:09:02.053 } 00:09:02.053 ], 00:09:02.053 "driver_specific": {} 00:09:02.053 } 00:09:02.053 ] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.053 [2024-10-29 10:58:07.519144] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:02.053 [2024-10-29 10:58:07.519295] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:02.053 [2024-10-29 10:58:07.519342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:02.053 [2024-10-29 10:58:07.521558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.053 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.313 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.313 "name": "Existed_Raid", 00:09:02.313 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.313 "strip_size_kb": 64, 00:09:02.313 "state": "configuring", 00:09:02.313 "raid_level": "concat", 00:09:02.313 "superblock": false, 00:09:02.313 "num_base_bdevs": 3, 00:09:02.313 "num_base_bdevs_discovered": 2, 00:09:02.313 "num_base_bdevs_operational": 3, 00:09:02.313 "base_bdevs_list": [ 00:09:02.313 { 00:09:02.313 "name": "BaseBdev1", 00:09:02.313 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.313 "is_configured": false, 00:09:02.313 "data_offset": 0, 00:09:02.313 "data_size": 0 00:09:02.313 }, 00:09:02.313 { 00:09:02.313 "name": "BaseBdev2", 00:09:02.313 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:02.313 "is_configured": true, 00:09:02.313 "data_offset": 0, 00:09:02.313 "data_size": 65536 00:09:02.313 }, 00:09:02.313 { 00:09:02.313 "name": "BaseBdev3", 00:09:02.313 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:02.313 "is_configured": true, 00:09:02.313 "data_offset": 0, 00:09:02.313 "data_size": 65536 00:09:02.313 } 00:09:02.313 ] 00:09:02.313 }' 00:09:02.313 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.313 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.573 [2024-10-29 10:58:07.906511] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.573 "name": "Existed_Raid", 00:09:02.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.573 "strip_size_kb": 64, 00:09:02.573 "state": "configuring", 00:09:02.573 "raid_level": "concat", 00:09:02.573 "superblock": false, 00:09:02.573 "num_base_bdevs": 3, 00:09:02.573 "num_base_bdevs_discovered": 1, 00:09:02.573 "num_base_bdevs_operational": 3, 00:09:02.573 "base_bdevs_list": [ 00:09:02.573 { 00:09:02.573 "name": "BaseBdev1", 00:09:02.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.573 "is_configured": false, 00:09:02.573 "data_offset": 0, 00:09:02.573 "data_size": 0 00:09:02.573 }, 00:09:02.573 { 00:09:02.573 "name": null, 00:09:02.573 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:02.573 "is_configured": false, 00:09:02.573 "data_offset": 0, 00:09:02.573 "data_size": 65536 00:09:02.573 }, 00:09:02.573 { 00:09:02.573 "name": "BaseBdev3", 00:09:02.573 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:02.573 "is_configured": true, 00:09:02.573 "data_offset": 0, 00:09:02.573 "data_size": 65536 00:09:02.573 } 00:09:02.573 ] 00:09:02.573 }' 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.573 10:58:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.143 [2024-10-29 10:58:08.422381] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:03.143 BaseBdev1 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.143 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.143 [ 00:09:03.143 { 00:09:03.143 "name": "BaseBdev1", 00:09:03.143 "aliases": [ 00:09:03.143 "3cebdaec-7fae-4397-8744-6347d678e528" 00:09:03.143 ], 00:09:03.143 "product_name": "Malloc disk", 00:09:03.143 "block_size": 512, 00:09:03.143 "num_blocks": 65536, 00:09:03.143 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:03.143 "assigned_rate_limits": { 00:09:03.143 "rw_ios_per_sec": 0, 00:09:03.143 "rw_mbytes_per_sec": 0, 00:09:03.143 "r_mbytes_per_sec": 0, 00:09:03.143 "w_mbytes_per_sec": 0 00:09:03.143 }, 00:09:03.143 "claimed": true, 00:09:03.143 "claim_type": "exclusive_write", 00:09:03.143 "zoned": false, 00:09:03.143 "supported_io_types": { 00:09:03.143 "read": true, 00:09:03.143 "write": true, 00:09:03.143 "unmap": true, 00:09:03.143 "flush": true, 00:09:03.143 "reset": true, 00:09:03.143 "nvme_admin": false, 00:09:03.143 "nvme_io": false, 00:09:03.143 "nvme_io_md": false, 00:09:03.143 "write_zeroes": true, 00:09:03.143 "zcopy": true, 00:09:03.143 "get_zone_info": false, 00:09:03.143 "zone_management": false, 00:09:03.143 "zone_append": false, 00:09:03.143 "compare": false, 00:09:03.143 "compare_and_write": false, 00:09:03.143 "abort": true, 00:09:03.143 "seek_hole": false, 00:09:03.143 "seek_data": false, 00:09:03.143 "copy": true, 00:09:03.143 "nvme_iov_md": false 00:09:03.143 }, 00:09:03.143 "memory_domains": [ 00:09:03.143 { 00:09:03.143 "dma_device_id": "system", 00:09:03.143 "dma_device_type": 1 00:09:03.143 }, 00:09:03.144 { 00:09:03.144 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.144 "dma_device_type": 2 00:09:03.144 } 00:09:03.144 ], 00:09:03.144 "driver_specific": {} 00:09:03.144 } 00:09:03.144 ] 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.144 "name": "Existed_Raid", 00:09:03.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.144 "strip_size_kb": 64, 00:09:03.144 "state": "configuring", 00:09:03.144 "raid_level": "concat", 00:09:03.144 "superblock": false, 00:09:03.144 "num_base_bdevs": 3, 00:09:03.144 "num_base_bdevs_discovered": 2, 00:09:03.144 "num_base_bdevs_operational": 3, 00:09:03.144 "base_bdevs_list": [ 00:09:03.144 { 00:09:03.144 "name": "BaseBdev1", 00:09:03.144 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:03.144 "is_configured": true, 00:09:03.144 "data_offset": 0, 00:09:03.144 "data_size": 65536 00:09:03.144 }, 00:09:03.144 { 00:09:03.144 "name": null, 00:09:03.144 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:03.144 "is_configured": false, 00:09:03.144 "data_offset": 0, 00:09:03.144 "data_size": 65536 00:09:03.144 }, 00:09:03.144 { 00:09:03.144 "name": "BaseBdev3", 00:09:03.144 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:03.144 "is_configured": true, 00:09:03.144 "data_offset": 0, 00:09:03.144 "data_size": 65536 00:09:03.144 } 00:09:03.144 ] 00:09:03.144 }' 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.144 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.720 [2024-10-29 10:58:08.977478] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.720 10:58:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.721 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.721 10:58:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.721 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.721 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.721 "name": "Existed_Raid", 00:09:03.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.721 "strip_size_kb": 64, 00:09:03.721 "state": "configuring", 00:09:03.721 "raid_level": "concat", 00:09:03.721 "superblock": false, 00:09:03.721 "num_base_bdevs": 3, 00:09:03.721 "num_base_bdevs_discovered": 1, 00:09:03.721 "num_base_bdevs_operational": 3, 00:09:03.721 "base_bdevs_list": [ 00:09:03.721 { 00:09:03.721 "name": "BaseBdev1", 00:09:03.721 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:03.721 "is_configured": true, 00:09:03.721 "data_offset": 0, 00:09:03.721 "data_size": 65536 00:09:03.721 }, 00:09:03.721 { 00:09:03.721 "name": null, 00:09:03.721 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:03.721 "is_configured": false, 00:09:03.721 "data_offset": 0, 00:09:03.721 "data_size": 65536 00:09:03.721 }, 00:09:03.721 { 00:09:03.721 "name": null, 00:09:03.721 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:03.721 "is_configured": false, 00:09:03.721 "data_offset": 0, 00:09:03.721 "data_size": 65536 00:09:03.721 } 00:09:03.721 ] 00:09:03.721 }' 00:09:03.721 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.721 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.999 [2024-10-29 10:58:09.428680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.999 "name": "Existed_Raid", 00:09:03.999 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.999 "strip_size_kb": 64, 00:09:03.999 "state": "configuring", 00:09:03.999 "raid_level": "concat", 00:09:03.999 "superblock": false, 00:09:03.999 "num_base_bdevs": 3, 00:09:03.999 "num_base_bdevs_discovered": 2, 00:09:03.999 "num_base_bdevs_operational": 3, 00:09:03.999 "base_bdevs_list": [ 00:09:03.999 { 00:09:03.999 "name": "BaseBdev1", 00:09:03.999 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:03.999 "is_configured": true, 00:09:03.999 "data_offset": 0, 00:09:03.999 "data_size": 65536 00:09:03.999 }, 00:09:03.999 { 00:09:03.999 "name": null, 00:09:03.999 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:03.999 "is_configured": false, 00:09:03.999 "data_offset": 0, 00:09:03.999 "data_size": 65536 00:09:03.999 }, 00:09:03.999 { 00:09:03.999 "name": "BaseBdev3", 00:09:03.999 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:03.999 "is_configured": true, 00:09:03.999 "data_offset": 0, 00:09:03.999 "data_size": 65536 00:09:03.999 } 00:09:03.999 ] 00:09:03.999 }' 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.999 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.651 [2024-10-29 10:58:09.891959] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.651 "name": "Existed_Raid", 00:09:04.651 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:04.651 "strip_size_kb": 64, 00:09:04.651 "state": "configuring", 00:09:04.651 "raid_level": "concat", 00:09:04.651 "superblock": false, 00:09:04.651 "num_base_bdevs": 3, 00:09:04.651 "num_base_bdevs_discovered": 1, 00:09:04.651 "num_base_bdevs_operational": 3, 00:09:04.651 "base_bdevs_list": [ 00:09:04.651 { 00:09:04.651 "name": null, 00:09:04.651 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:04.651 "is_configured": false, 00:09:04.651 "data_offset": 0, 00:09:04.651 "data_size": 65536 00:09:04.651 }, 00:09:04.651 { 00:09:04.651 "name": null, 00:09:04.651 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:04.651 "is_configured": false, 00:09:04.651 "data_offset": 0, 00:09:04.651 "data_size": 65536 00:09:04.651 }, 00:09:04.651 { 00:09:04.651 "name": "BaseBdev3", 00:09:04.651 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:04.651 "is_configured": true, 00:09:04.651 "data_offset": 0, 00:09:04.651 "data_size": 65536 00:09:04.651 } 00:09:04.651 ] 00:09:04.651 }' 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.651 10:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.911 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.911 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.911 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.911 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:04.912 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.912 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:04.912 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:04.912 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.912 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.912 [2024-10-29 10:58:10.411251] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.172 "name": "Existed_Raid", 00:09:05.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.172 "strip_size_kb": 64, 00:09:05.172 "state": "configuring", 00:09:05.172 "raid_level": "concat", 00:09:05.172 "superblock": false, 00:09:05.172 "num_base_bdevs": 3, 00:09:05.172 "num_base_bdevs_discovered": 2, 00:09:05.172 "num_base_bdevs_operational": 3, 00:09:05.172 "base_bdevs_list": [ 00:09:05.172 { 00:09:05.172 "name": null, 00:09:05.172 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:05.172 "is_configured": false, 00:09:05.172 "data_offset": 0, 00:09:05.172 "data_size": 65536 00:09:05.172 }, 00:09:05.172 { 00:09:05.172 "name": "BaseBdev2", 00:09:05.172 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:05.172 "is_configured": true, 00:09:05.172 "data_offset": 0, 00:09:05.172 "data_size": 65536 00:09:05.172 }, 00:09:05.172 { 00:09:05.172 "name": "BaseBdev3", 00:09:05.172 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:05.172 "is_configured": true, 00:09:05.172 "data_offset": 0, 00:09:05.172 "data_size": 65536 00:09:05.172 } 00:09:05.172 ] 00:09:05.172 }' 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.172 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:05.432 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3cebdaec-7fae-4397-8744-6347d678e528 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.692 [2024-10-29 10:58:10.987264] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:05.692 [2024-10-29 10:58:10.987408] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:05.692 [2024-10-29 10:58:10.987433] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:05.692 [2024-10-29 10:58:10.987775] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:05.692 [2024-10-29 10:58:10.987914] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:05.692 [2024-10-29 10:58:10.987923] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:05.692 [2024-10-29 10:58:10.988145] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.692 NewBaseBdev 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.692 10:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.692 [ 00:09:05.692 { 00:09:05.692 "name": "NewBaseBdev", 00:09:05.692 "aliases": [ 00:09:05.692 "3cebdaec-7fae-4397-8744-6347d678e528" 00:09:05.692 ], 00:09:05.692 "product_name": "Malloc disk", 00:09:05.692 "block_size": 512, 00:09:05.692 "num_blocks": 65536, 00:09:05.692 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:05.692 "assigned_rate_limits": { 00:09:05.692 "rw_ios_per_sec": 0, 00:09:05.692 "rw_mbytes_per_sec": 0, 00:09:05.692 "r_mbytes_per_sec": 0, 00:09:05.692 "w_mbytes_per_sec": 0 00:09:05.692 }, 00:09:05.692 "claimed": true, 00:09:05.692 "claim_type": "exclusive_write", 00:09:05.692 "zoned": false, 00:09:05.692 "supported_io_types": { 00:09:05.692 "read": true, 00:09:05.692 "write": true, 00:09:05.692 "unmap": true, 00:09:05.692 "flush": true, 00:09:05.692 "reset": true, 00:09:05.692 "nvme_admin": false, 00:09:05.692 "nvme_io": false, 00:09:05.692 "nvme_io_md": false, 00:09:05.692 "write_zeroes": true, 00:09:05.692 "zcopy": true, 00:09:05.692 "get_zone_info": false, 00:09:05.692 "zone_management": false, 00:09:05.693 "zone_append": false, 00:09:05.693 "compare": false, 00:09:05.693 "compare_and_write": false, 00:09:05.693 "abort": true, 00:09:05.693 "seek_hole": false, 00:09:05.693 "seek_data": false, 00:09:05.693 "copy": true, 00:09:05.693 "nvme_iov_md": false 00:09:05.693 }, 00:09:05.693 "memory_domains": [ 00:09:05.693 { 00:09:05.693 "dma_device_id": "system", 00:09:05.693 "dma_device_type": 1 00:09:05.693 }, 00:09:05.693 { 00:09:05.693 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.693 "dma_device_type": 2 00:09:05.693 } 00:09:05.693 ], 00:09:05.693 "driver_specific": {} 00:09:05.693 } 00:09:05.693 ] 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.693 "name": "Existed_Raid", 00:09:05.693 "uuid": "6eb0a8f7-9872-4272-9581-853965f744ca", 00:09:05.693 "strip_size_kb": 64, 00:09:05.693 "state": "online", 00:09:05.693 "raid_level": "concat", 00:09:05.693 "superblock": false, 00:09:05.693 "num_base_bdevs": 3, 00:09:05.693 "num_base_bdevs_discovered": 3, 00:09:05.693 "num_base_bdevs_operational": 3, 00:09:05.693 "base_bdevs_list": [ 00:09:05.693 { 00:09:05.693 "name": "NewBaseBdev", 00:09:05.693 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:05.693 "is_configured": true, 00:09:05.693 "data_offset": 0, 00:09:05.693 "data_size": 65536 00:09:05.693 }, 00:09:05.693 { 00:09:05.693 "name": "BaseBdev2", 00:09:05.693 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:05.693 "is_configured": true, 00:09:05.693 "data_offset": 0, 00:09:05.693 "data_size": 65536 00:09:05.693 }, 00:09:05.693 { 00:09:05.693 "name": "BaseBdev3", 00:09:05.693 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:05.693 "is_configured": true, 00:09:05.693 "data_offset": 0, 00:09:05.693 "data_size": 65536 00:09:05.693 } 00:09:05.693 ] 00:09:05.693 }' 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.693 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.264 [2024-10-29 10:58:11.530794] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:06.264 "name": "Existed_Raid", 00:09:06.264 "aliases": [ 00:09:06.264 "6eb0a8f7-9872-4272-9581-853965f744ca" 00:09:06.264 ], 00:09:06.264 "product_name": "Raid Volume", 00:09:06.264 "block_size": 512, 00:09:06.264 "num_blocks": 196608, 00:09:06.264 "uuid": "6eb0a8f7-9872-4272-9581-853965f744ca", 00:09:06.264 "assigned_rate_limits": { 00:09:06.264 "rw_ios_per_sec": 0, 00:09:06.264 "rw_mbytes_per_sec": 0, 00:09:06.264 "r_mbytes_per_sec": 0, 00:09:06.264 "w_mbytes_per_sec": 0 00:09:06.264 }, 00:09:06.264 "claimed": false, 00:09:06.264 "zoned": false, 00:09:06.264 "supported_io_types": { 00:09:06.264 "read": true, 00:09:06.264 "write": true, 00:09:06.264 "unmap": true, 00:09:06.264 "flush": true, 00:09:06.264 "reset": true, 00:09:06.264 "nvme_admin": false, 00:09:06.264 "nvme_io": false, 00:09:06.264 "nvme_io_md": false, 00:09:06.264 "write_zeroes": true, 00:09:06.264 "zcopy": false, 00:09:06.264 "get_zone_info": false, 00:09:06.264 "zone_management": false, 00:09:06.264 "zone_append": false, 00:09:06.264 "compare": false, 00:09:06.264 "compare_and_write": false, 00:09:06.264 "abort": false, 00:09:06.264 "seek_hole": false, 00:09:06.264 "seek_data": false, 00:09:06.264 "copy": false, 00:09:06.264 "nvme_iov_md": false 00:09:06.264 }, 00:09:06.264 "memory_domains": [ 00:09:06.264 { 00:09:06.264 "dma_device_id": "system", 00:09:06.264 "dma_device_type": 1 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.264 "dma_device_type": 2 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "dma_device_id": "system", 00:09:06.264 "dma_device_type": 1 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.264 "dma_device_type": 2 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "dma_device_id": "system", 00:09:06.264 "dma_device_type": 1 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.264 "dma_device_type": 2 00:09:06.264 } 00:09:06.264 ], 00:09:06.264 "driver_specific": { 00:09:06.264 "raid": { 00:09:06.264 "uuid": "6eb0a8f7-9872-4272-9581-853965f744ca", 00:09:06.264 "strip_size_kb": 64, 00:09:06.264 "state": "online", 00:09:06.264 "raid_level": "concat", 00:09:06.264 "superblock": false, 00:09:06.264 "num_base_bdevs": 3, 00:09:06.264 "num_base_bdevs_discovered": 3, 00:09:06.264 "num_base_bdevs_operational": 3, 00:09:06.264 "base_bdevs_list": [ 00:09:06.264 { 00:09:06.264 "name": "NewBaseBdev", 00:09:06.264 "uuid": "3cebdaec-7fae-4397-8744-6347d678e528", 00:09:06.264 "is_configured": true, 00:09:06.264 "data_offset": 0, 00:09:06.264 "data_size": 65536 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "name": "BaseBdev2", 00:09:06.264 "uuid": "26c1c4ff-004c-420c-8833-a94bd8240128", 00:09:06.264 "is_configured": true, 00:09:06.264 "data_offset": 0, 00:09:06.264 "data_size": 65536 00:09:06.264 }, 00:09:06.264 { 00:09:06.264 "name": "BaseBdev3", 00:09:06.264 "uuid": "ef65c798-9cab-4425-ae0b-57b08b1616a9", 00:09:06.264 "is_configured": true, 00:09:06.264 "data_offset": 0, 00:09:06.264 "data_size": 65536 00:09:06.264 } 00:09:06.264 ] 00:09:06.264 } 00:09:06.264 } 00:09:06.264 }' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:06.264 BaseBdev2 00:09:06.264 BaseBdev3' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.264 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.526 [2024-10-29 10:58:11.805950] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:06.526 [2024-10-29 10:58:11.806062] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:06.526 [2024-10-29 10:58:11.806188] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:06.526 [2024-10-29 10:58:11.806294] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:06.526 [2024-10-29 10:58:11.806354] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76882 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 76882 ']' 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 76882 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 76882 00:09:06.526 killing process with pid 76882 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 76882' 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 76882 00:09:06.526 [2024-10-29 10:58:11.851766] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:06.526 10:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 76882 00:09:06.526 [2024-10-29 10:58:11.912571] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:06.786 10:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:06.786 00:09:06.786 real 0m9.036s 00:09:06.786 user 0m15.156s 00:09:06.786 sys 0m1.892s 00:09:06.786 10:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:06.786 10:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.786 ************************************ 00:09:06.786 END TEST raid_state_function_test 00:09:06.786 ************************************ 00:09:07.047 10:58:12 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:09:07.047 10:58:12 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:07.047 10:58:12 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:07.047 10:58:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:07.047 ************************************ 00:09:07.047 START TEST raid_state_function_test_sb 00:09:07.047 ************************************ 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 3 true 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:07.047 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77492 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77492' 00:09:07.048 Process raid pid: 77492 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77492 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 77492 ']' 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:07.048 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:07.048 10:58:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.048 [2024-10-29 10:58:12.410951] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:07.048 [2024-10-29 10:58:12.411203] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:07.308 [2024-10-29 10:58:12.585349] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.308 [2024-10-29 10:58:12.630187] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.308 [2024-10-29 10:58:12.707342] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.308 [2024-10-29 10:58:12.707500] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.878 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:07.878 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.879 [2024-10-29 10:58:13.227477] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:07.879 [2024-10-29 10:58:13.227926] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:07.879 [2024-10-29 10:58:13.227960] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:07.879 [2024-10-29 10:58:13.228022] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:07.879 [2024-10-29 10:58:13.228033] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:07.879 [2024-10-29 10:58:13.228097] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.879 "name": "Existed_Raid", 00:09:07.879 "uuid": "e8b613aa-2330-4e74-902c-e8adb7ef1744", 00:09:07.879 "strip_size_kb": 64, 00:09:07.879 "state": "configuring", 00:09:07.879 "raid_level": "concat", 00:09:07.879 "superblock": true, 00:09:07.879 "num_base_bdevs": 3, 00:09:07.879 "num_base_bdevs_discovered": 0, 00:09:07.879 "num_base_bdevs_operational": 3, 00:09:07.879 "base_bdevs_list": [ 00:09:07.879 { 00:09:07.879 "name": "BaseBdev1", 00:09:07.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.879 "is_configured": false, 00:09:07.879 "data_offset": 0, 00:09:07.879 "data_size": 0 00:09:07.879 }, 00:09:07.879 { 00:09:07.879 "name": "BaseBdev2", 00:09:07.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.879 "is_configured": false, 00:09:07.879 "data_offset": 0, 00:09:07.879 "data_size": 0 00:09:07.879 }, 00:09:07.879 { 00:09:07.879 "name": "BaseBdev3", 00:09:07.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.879 "is_configured": false, 00:09:07.879 "data_offset": 0, 00:09:07.879 "data_size": 0 00:09:07.879 } 00:09:07.879 ] 00:09:07.879 }' 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.879 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.139 [2024-10-29 10:58:13.570758] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:08.139 [2024-10-29 10:58:13.570903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.139 [2024-10-29 10:58:13.582764] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:08.139 [2024-10-29 10:58:13.583352] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:08.139 [2024-10-29 10:58:13.583456] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.139 [2024-10-29 10:58:13.583574] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.139 [2024-10-29 10:58:13.583613] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.139 [2024-10-29 10:58:13.583701] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.139 [2024-10-29 10:58:13.610102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.139 BaseBdev1 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.139 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.399 [ 00:09:08.399 { 00:09:08.399 "name": "BaseBdev1", 00:09:08.399 "aliases": [ 00:09:08.399 "c7ea1dff-8a93-464c-999c-cd994cf6be24" 00:09:08.399 ], 00:09:08.399 "product_name": "Malloc disk", 00:09:08.399 "block_size": 512, 00:09:08.400 "num_blocks": 65536, 00:09:08.400 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:08.400 "assigned_rate_limits": { 00:09:08.400 "rw_ios_per_sec": 0, 00:09:08.400 "rw_mbytes_per_sec": 0, 00:09:08.400 "r_mbytes_per_sec": 0, 00:09:08.400 "w_mbytes_per_sec": 0 00:09:08.400 }, 00:09:08.400 "claimed": true, 00:09:08.400 "claim_type": "exclusive_write", 00:09:08.400 "zoned": false, 00:09:08.400 "supported_io_types": { 00:09:08.400 "read": true, 00:09:08.400 "write": true, 00:09:08.400 "unmap": true, 00:09:08.400 "flush": true, 00:09:08.400 "reset": true, 00:09:08.400 "nvme_admin": false, 00:09:08.400 "nvme_io": false, 00:09:08.400 "nvme_io_md": false, 00:09:08.400 "write_zeroes": true, 00:09:08.400 "zcopy": true, 00:09:08.400 "get_zone_info": false, 00:09:08.400 "zone_management": false, 00:09:08.400 "zone_append": false, 00:09:08.400 "compare": false, 00:09:08.400 "compare_and_write": false, 00:09:08.400 "abort": true, 00:09:08.400 "seek_hole": false, 00:09:08.400 "seek_data": false, 00:09:08.400 "copy": true, 00:09:08.400 "nvme_iov_md": false 00:09:08.400 }, 00:09:08.400 "memory_domains": [ 00:09:08.400 { 00:09:08.400 "dma_device_id": "system", 00:09:08.400 "dma_device_type": 1 00:09:08.400 }, 00:09:08.400 { 00:09:08.400 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.400 "dma_device_type": 2 00:09:08.400 } 00:09:08.400 ], 00:09:08.400 "driver_specific": {} 00:09:08.400 } 00:09:08.400 ] 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.400 "name": "Existed_Raid", 00:09:08.400 "uuid": "88a59b21-e086-4102-8b2b-0ad27bf7ed53", 00:09:08.400 "strip_size_kb": 64, 00:09:08.400 "state": "configuring", 00:09:08.400 "raid_level": "concat", 00:09:08.400 "superblock": true, 00:09:08.400 "num_base_bdevs": 3, 00:09:08.400 "num_base_bdevs_discovered": 1, 00:09:08.400 "num_base_bdevs_operational": 3, 00:09:08.400 "base_bdevs_list": [ 00:09:08.400 { 00:09:08.400 "name": "BaseBdev1", 00:09:08.400 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:08.400 "is_configured": true, 00:09:08.400 "data_offset": 2048, 00:09:08.400 "data_size": 63488 00:09:08.400 }, 00:09:08.400 { 00:09:08.400 "name": "BaseBdev2", 00:09:08.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.400 "is_configured": false, 00:09:08.400 "data_offset": 0, 00:09:08.400 "data_size": 0 00:09:08.400 }, 00:09:08.400 { 00:09:08.400 "name": "BaseBdev3", 00:09:08.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.400 "is_configured": false, 00:09:08.400 "data_offset": 0, 00:09:08.400 "data_size": 0 00:09:08.400 } 00:09:08.400 ] 00:09:08.400 }' 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.400 10:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.661 [2024-10-29 10:58:14.049472] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:08.661 [2024-10-29 10:58:14.049619] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.661 [2024-10-29 10:58:14.061487] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.661 [2024-10-29 10:58:14.063728] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.661 [2024-10-29 10:58:14.063992] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.661 [2024-10-29 10:58:14.064010] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.661 [2024-10-29 10:58:14.064073] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.661 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.661 "name": "Existed_Raid", 00:09:08.661 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:08.661 "strip_size_kb": 64, 00:09:08.661 "state": "configuring", 00:09:08.661 "raid_level": "concat", 00:09:08.661 "superblock": true, 00:09:08.661 "num_base_bdevs": 3, 00:09:08.661 "num_base_bdevs_discovered": 1, 00:09:08.661 "num_base_bdevs_operational": 3, 00:09:08.661 "base_bdevs_list": [ 00:09:08.661 { 00:09:08.661 "name": "BaseBdev1", 00:09:08.661 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:08.661 "is_configured": true, 00:09:08.661 "data_offset": 2048, 00:09:08.661 "data_size": 63488 00:09:08.661 }, 00:09:08.661 { 00:09:08.661 "name": "BaseBdev2", 00:09:08.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.661 "is_configured": false, 00:09:08.661 "data_offset": 0, 00:09:08.661 "data_size": 0 00:09:08.661 }, 00:09:08.661 { 00:09:08.661 "name": "BaseBdev3", 00:09:08.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.661 "is_configured": false, 00:09:08.661 "data_offset": 0, 00:09:08.661 "data_size": 0 00:09:08.661 } 00:09:08.661 ] 00:09:08.661 }' 00:09:08.662 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.662 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.233 [2024-10-29 10:58:14.509543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:09.233 BaseBdev2 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.233 [ 00:09:09.233 { 00:09:09.233 "name": "BaseBdev2", 00:09:09.233 "aliases": [ 00:09:09.233 "e4281be6-7a87-4a81-91d3-1a0ae5283282" 00:09:09.233 ], 00:09:09.233 "product_name": "Malloc disk", 00:09:09.233 "block_size": 512, 00:09:09.233 "num_blocks": 65536, 00:09:09.233 "uuid": "e4281be6-7a87-4a81-91d3-1a0ae5283282", 00:09:09.233 "assigned_rate_limits": { 00:09:09.233 "rw_ios_per_sec": 0, 00:09:09.233 "rw_mbytes_per_sec": 0, 00:09:09.233 "r_mbytes_per_sec": 0, 00:09:09.233 "w_mbytes_per_sec": 0 00:09:09.233 }, 00:09:09.233 "claimed": true, 00:09:09.233 "claim_type": "exclusive_write", 00:09:09.233 "zoned": false, 00:09:09.233 "supported_io_types": { 00:09:09.233 "read": true, 00:09:09.233 "write": true, 00:09:09.233 "unmap": true, 00:09:09.233 "flush": true, 00:09:09.233 "reset": true, 00:09:09.233 "nvme_admin": false, 00:09:09.233 "nvme_io": false, 00:09:09.233 "nvme_io_md": false, 00:09:09.233 "write_zeroes": true, 00:09:09.233 "zcopy": true, 00:09:09.233 "get_zone_info": false, 00:09:09.233 "zone_management": false, 00:09:09.233 "zone_append": false, 00:09:09.233 "compare": false, 00:09:09.233 "compare_and_write": false, 00:09:09.233 "abort": true, 00:09:09.233 "seek_hole": false, 00:09:09.233 "seek_data": false, 00:09:09.233 "copy": true, 00:09:09.233 "nvme_iov_md": false 00:09:09.233 }, 00:09:09.233 "memory_domains": [ 00:09:09.233 { 00:09:09.233 "dma_device_id": "system", 00:09:09.233 "dma_device_type": 1 00:09:09.233 }, 00:09:09.233 { 00:09:09.233 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.233 "dma_device_type": 2 00:09:09.233 } 00:09:09.233 ], 00:09:09.233 "driver_specific": {} 00:09:09.233 } 00:09:09.233 ] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.233 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.233 "name": "Existed_Raid", 00:09:09.233 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:09.233 "strip_size_kb": 64, 00:09:09.233 "state": "configuring", 00:09:09.233 "raid_level": "concat", 00:09:09.233 "superblock": true, 00:09:09.233 "num_base_bdevs": 3, 00:09:09.233 "num_base_bdevs_discovered": 2, 00:09:09.233 "num_base_bdevs_operational": 3, 00:09:09.233 "base_bdevs_list": [ 00:09:09.233 { 00:09:09.233 "name": "BaseBdev1", 00:09:09.233 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:09.233 "is_configured": true, 00:09:09.233 "data_offset": 2048, 00:09:09.233 "data_size": 63488 00:09:09.233 }, 00:09:09.233 { 00:09:09.233 "name": "BaseBdev2", 00:09:09.233 "uuid": "e4281be6-7a87-4a81-91d3-1a0ae5283282", 00:09:09.233 "is_configured": true, 00:09:09.233 "data_offset": 2048, 00:09:09.233 "data_size": 63488 00:09:09.233 }, 00:09:09.233 { 00:09:09.233 "name": "BaseBdev3", 00:09:09.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.233 "is_configured": false, 00:09:09.233 "data_offset": 0, 00:09:09.233 "data_size": 0 00:09:09.234 } 00:09:09.234 ] 00:09:09.234 }' 00:09:09.234 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.234 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 [2024-10-29 10:58:14.967671] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:09.494 [2024-10-29 10:58:14.967933] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:09.494 [2024-10-29 10:58:14.967957] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:09.494 [2024-10-29 10:58:14.968321] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:09.494 BaseBdev3 00:09:09.494 [2024-10-29 10:58:14.968550] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:09.494 [2024-10-29 10:58:14.968566] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:09.494 [2024-10-29 10:58:14.968755] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.494 10:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.494 [ 00:09:09.494 { 00:09:09.494 "name": "BaseBdev3", 00:09:09.494 "aliases": [ 00:09:09.494 "6eb11f04-17ff-459c-b06f-35b01a294e45" 00:09:09.754 ], 00:09:09.754 "product_name": "Malloc disk", 00:09:09.754 "block_size": 512, 00:09:09.754 "num_blocks": 65536, 00:09:09.754 "uuid": "6eb11f04-17ff-459c-b06f-35b01a294e45", 00:09:09.754 "assigned_rate_limits": { 00:09:09.754 "rw_ios_per_sec": 0, 00:09:09.754 "rw_mbytes_per_sec": 0, 00:09:09.754 "r_mbytes_per_sec": 0, 00:09:09.754 "w_mbytes_per_sec": 0 00:09:09.754 }, 00:09:09.754 "claimed": true, 00:09:09.754 "claim_type": "exclusive_write", 00:09:09.754 "zoned": false, 00:09:09.754 "supported_io_types": { 00:09:09.754 "read": true, 00:09:09.754 "write": true, 00:09:09.754 "unmap": true, 00:09:09.754 "flush": true, 00:09:09.754 "reset": true, 00:09:09.754 "nvme_admin": false, 00:09:09.754 "nvme_io": false, 00:09:09.754 "nvme_io_md": false, 00:09:09.754 "write_zeroes": true, 00:09:09.754 "zcopy": true, 00:09:09.754 "get_zone_info": false, 00:09:09.754 "zone_management": false, 00:09:09.754 "zone_append": false, 00:09:09.754 "compare": false, 00:09:09.754 "compare_and_write": false, 00:09:09.754 "abort": true, 00:09:09.754 "seek_hole": false, 00:09:09.754 "seek_data": false, 00:09:09.754 "copy": true, 00:09:09.754 "nvme_iov_md": false 00:09:09.754 }, 00:09:09.754 "memory_domains": [ 00:09:09.754 { 00:09:09.754 "dma_device_id": "system", 00:09:09.754 "dma_device_type": 1 00:09:09.754 }, 00:09:09.754 { 00:09:09.754 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.754 "dma_device_type": 2 00:09:09.754 } 00:09:09.754 ], 00:09:09.754 "driver_specific": {} 00:09:09.754 } 00:09:09.754 ] 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.754 "name": "Existed_Raid", 00:09:09.754 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:09.754 "strip_size_kb": 64, 00:09:09.754 "state": "online", 00:09:09.754 "raid_level": "concat", 00:09:09.754 "superblock": true, 00:09:09.754 "num_base_bdevs": 3, 00:09:09.754 "num_base_bdevs_discovered": 3, 00:09:09.754 "num_base_bdevs_operational": 3, 00:09:09.754 "base_bdevs_list": [ 00:09:09.754 { 00:09:09.754 "name": "BaseBdev1", 00:09:09.754 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:09.754 "is_configured": true, 00:09:09.754 "data_offset": 2048, 00:09:09.754 "data_size": 63488 00:09:09.754 }, 00:09:09.754 { 00:09:09.754 "name": "BaseBdev2", 00:09:09.754 "uuid": "e4281be6-7a87-4a81-91d3-1a0ae5283282", 00:09:09.754 "is_configured": true, 00:09:09.754 "data_offset": 2048, 00:09:09.754 "data_size": 63488 00:09:09.754 }, 00:09:09.754 { 00:09:09.754 "name": "BaseBdev3", 00:09:09.754 "uuid": "6eb11f04-17ff-459c-b06f-35b01a294e45", 00:09:09.754 "is_configured": true, 00:09:09.754 "data_offset": 2048, 00:09:09.754 "data_size": 63488 00:09:09.754 } 00:09:09.754 ] 00:09:09.754 }' 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.754 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:10.014 [2024-10-29 10:58:15.451399] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.014 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:10.014 "name": "Existed_Raid", 00:09:10.014 "aliases": [ 00:09:10.014 "2340cf96-6855-4d9c-aa2d-e80195a493d4" 00:09:10.014 ], 00:09:10.014 "product_name": "Raid Volume", 00:09:10.014 "block_size": 512, 00:09:10.014 "num_blocks": 190464, 00:09:10.014 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:10.014 "assigned_rate_limits": { 00:09:10.014 "rw_ios_per_sec": 0, 00:09:10.014 "rw_mbytes_per_sec": 0, 00:09:10.014 "r_mbytes_per_sec": 0, 00:09:10.014 "w_mbytes_per_sec": 0 00:09:10.014 }, 00:09:10.014 "claimed": false, 00:09:10.014 "zoned": false, 00:09:10.014 "supported_io_types": { 00:09:10.014 "read": true, 00:09:10.014 "write": true, 00:09:10.014 "unmap": true, 00:09:10.014 "flush": true, 00:09:10.014 "reset": true, 00:09:10.014 "nvme_admin": false, 00:09:10.014 "nvme_io": false, 00:09:10.014 "nvme_io_md": false, 00:09:10.014 "write_zeroes": true, 00:09:10.014 "zcopy": false, 00:09:10.014 "get_zone_info": false, 00:09:10.014 "zone_management": false, 00:09:10.014 "zone_append": false, 00:09:10.014 "compare": false, 00:09:10.014 "compare_and_write": false, 00:09:10.014 "abort": false, 00:09:10.014 "seek_hole": false, 00:09:10.014 "seek_data": false, 00:09:10.014 "copy": false, 00:09:10.014 "nvme_iov_md": false 00:09:10.014 }, 00:09:10.015 "memory_domains": [ 00:09:10.015 { 00:09:10.015 "dma_device_id": "system", 00:09:10.015 "dma_device_type": 1 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.015 "dma_device_type": 2 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "dma_device_id": "system", 00:09:10.015 "dma_device_type": 1 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.015 "dma_device_type": 2 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "dma_device_id": "system", 00:09:10.015 "dma_device_type": 1 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.015 "dma_device_type": 2 00:09:10.015 } 00:09:10.015 ], 00:09:10.015 "driver_specific": { 00:09:10.015 "raid": { 00:09:10.015 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:10.015 "strip_size_kb": 64, 00:09:10.015 "state": "online", 00:09:10.015 "raid_level": "concat", 00:09:10.015 "superblock": true, 00:09:10.015 "num_base_bdevs": 3, 00:09:10.015 "num_base_bdevs_discovered": 3, 00:09:10.015 "num_base_bdevs_operational": 3, 00:09:10.015 "base_bdevs_list": [ 00:09:10.015 { 00:09:10.015 "name": "BaseBdev1", 00:09:10.015 "uuid": "c7ea1dff-8a93-464c-999c-cd994cf6be24", 00:09:10.015 "is_configured": true, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "name": "BaseBdev2", 00:09:10.015 "uuid": "e4281be6-7a87-4a81-91d3-1a0ae5283282", 00:09:10.015 "is_configured": true, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 }, 00:09:10.015 { 00:09:10.015 "name": "BaseBdev3", 00:09:10.015 "uuid": "6eb11f04-17ff-459c-b06f-35b01a294e45", 00:09:10.015 "is_configured": true, 00:09:10.015 "data_offset": 2048, 00:09:10.015 "data_size": 63488 00:09:10.015 } 00:09:10.015 ] 00:09:10.015 } 00:09:10.015 } 00:09:10.015 }' 00:09:10.015 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:10.276 BaseBdev2 00:09:10.276 BaseBdev3' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.276 [2024-10-29 10:58:15.678657] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:10.276 [2024-10-29 10:58:15.678775] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:10.276 [2024-10-29 10:58:15.678845] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.276 "name": "Existed_Raid", 00:09:10.276 "uuid": "2340cf96-6855-4d9c-aa2d-e80195a493d4", 00:09:10.276 "strip_size_kb": 64, 00:09:10.276 "state": "offline", 00:09:10.276 "raid_level": "concat", 00:09:10.276 "superblock": true, 00:09:10.276 "num_base_bdevs": 3, 00:09:10.276 "num_base_bdevs_discovered": 2, 00:09:10.276 "num_base_bdevs_operational": 2, 00:09:10.276 "base_bdevs_list": [ 00:09:10.276 { 00:09:10.276 "name": null, 00:09:10.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:10.276 "is_configured": false, 00:09:10.276 "data_offset": 0, 00:09:10.276 "data_size": 63488 00:09:10.276 }, 00:09:10.276 { 00:09:10.276 "name": "BaseBdev2", 00:09:10.276 "uuid": "e4281be6-7a87-4a81-91d3-1a0ae5283282", 00:09:10.276 "is_configured": true, 00:09:10.276 "data_offset": 2048, 00:09:10.276 "data_size": 63488 00:09:10.276 }, 00:09:10.276 { 00:09:10.276 "name": "BaseBdev3", 00:09:10.276 "uuid": "6eb11f04-17ff-459c-b06f-35b01a294e45", 00:09:10.276 "is_configured": true, 00:09:10.276 "data_offset": 2048, 00:09:10.276 "data_size": 63488 00:09:10.276 } 00:09:10.276 ] 00:09:10.276 }' 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.276 10:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 [2024-10-29 10:58:16.234843] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 [2024-10-29 10:58:16.315348] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:10.848 [2024-10-29 10:58:16.315425] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.848 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.109 BaseBdev2 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.109 [ 00:09:11.109 { 00:09:11.109 "name": "BaseBdev2", 00:09:11.109 "aliases": [ 00:09:11.109 "aa88c035-5474-4dae-a4ee-6e15e42a03be" 00:09:11.109 ], 00:09:11.109 "product_name": "Malloc disk", 00:09:11.109 "block_size": 512, 00:09:11.109 "num_blocks": 65536, 00:09:11.109 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:11.109 "assigned_rate_limits": { 00:09:11.109 "rw_ios_per_sec": 0, 00:09:11.109 "rw_mbytes_per_sec": 0, 00:09:11.109 "r_mbytes_per_sec": 0, 00:09:11.109 "w_mbytes_per_sec": 0 00:09:11.109 }, 00:09:11.109 "claimed": false, 00:09:11.109 "zoned": false, 00:09:11.109 "supported_io_types": { 00:09:11.109 "read": true, 00:09:11.109 "write": true, 00:09:11.109 "unmap": true, 00:09:11.109 "flush": true, 00:09:11.109 "reset": true, 00:09:11.109 "nvme_admin": false, 00:09:11.109 "nvme_io": false, 00:09:11.109 "nvme_io_md": false, 00:09:11.109 "write_zeroes": true, 00:09:11.109 "zcopy": true, 00:09:11.109 "get_zone_info": false, 00:09:11.109 "zone_management": false, 00:09:11.109 "zone_append": false, 00:09:11.109 "compare": false, 00:09:11.109 "compare_and_write": false, 00:09:11.109 "abort": true, 00:09:11.109 "seek_hole": false, 00:09:11.109 "seek_data": false, 00:09:11.109 "copy": true, 00:09:11.109 "nvme_iov_md": false 00:09:11.109 }, 00:09:11.109 "memory_domains": [ 00:09:11.109 { 00:09:11.109 "dma_device_id": "system", 00:09:11.109 "dma_device_type": 1 00:09:11.109 }, 00:09:11.109 { 00:09:11.109 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.109 "dma_device_type": 2 00:09:11.109 } 00:09:11.109 ], 00:09:11.109 "driver_specific": {} 00:09:11.109 } 00:09:11.109 ] 00:09:11.109 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.110 BaseBdev3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.110 [ 00:09:11.110 { 00:09:11.110 "name": "BaseBdev3", 00:09:11.110 "aliases": [ 00:09:11.110 "3bbb7d66-56c9-4adc-a3f0-07c1df99242c" 00:09:11.110 ], 00:09:11.110 "product_name": "Malloc disk", 00:09:11.110 "block_size": 512, 00:09:11.110 "num_blocks": 65536, 00:09:11.110 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:11.110 "assigned_rate_limits": { 00:09:11.110 "rw_ios_per_sec": 0, 00:09:11.110 "rw_mbytes_per_sec": 0, 00:09:11.110 "r_mbytes_per_sec": 0, 00:09:11.110 "w_mbytes_per_sec": 0 00:09:11.110 }, 00:09:11.110 "claimed": false, 00:09:11.110 "zoned": false, 00:09:11.110 "supported_io_types": { 00:09:11.110 "read": true, 00:09:11.110 "write": true, 00:09:11.110 "unmap": true, 00:09:11.110 "flush": true, 00:09:11.110 "reset": true, 00:09:11.110 "nvme_admin": false, 00:09:11.110 "nvme_io": false, 00:09:11.110 "nvme_io_md": false, 00:09:11.110 "write_zeroes": true, 00:09:11.110 "zcopy": true, 00:09:11.110 "get_zone_info": false, 00:09:11.110 "zone_management": false, 00:09:11.110 "zone_append": false, 00:09:11.110 "compare": false, 00:09:11.110 "compare_and_write": false, 00:09:11.110 "abort": true, 00:09:11.110 "seek_hole": false, 00:09:11.110 "seek_data": false, 00:09:11.110 "copy": true, 00:09:11.110 "nvme_iov_md": false 00:09:11.110 }, 00:09:11.110 "memory_domains": [ 00:09:11.110 { 00:09:11.110 "dma_device_id": "system", 00:09:11.110 "dma_device_type": 1 00:09:11.110 }, 00:09:11.110 { 00:09:11.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.110 "dma_device_type": 2 00:09:11.110 } 00:09:11.110 ], 00:09:11.110 "driver_specific": {} 00:09:11.110 } 00:09:11.110 ] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.110 [2024-10-29 10:58:16.512161] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:11.110 [2024-10-29 10:58:16.512289] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:11.110 [2024-10-29 10:58:16.512335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.110 [2024-10-29 10:58:16.514496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.110 "name": "Existed_Raid", 00:09:11.110 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:11.110 "strip_size_kb": 64, 00:09:11.110 "state": "configuring", 00:09:11.110 "raid_level": "concat", 00:09:11.110 "superblock": true, 00:09:11.110 "num_base_bdevs": 3, 00:09:11.110 "num_base_bdevs_discovered": 2, 00:09:11.110 "num_base_bdevs_operational": 3, 00:09:11.110 "base_bdevs_list": [ 00:09:11.110 { 00:09:11.110 "name": "BaseBdev1", 00:09:11.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.110 "is_configured": false, 00:09:11.110 "data_offset": 0, 00:09:11.110 "data_size": 0 00:09:11.110 }, 00:09:11.110 { 00:09:11.110 "name": "BaseBdev2", 00:09:11.110 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:11.110 "is_configured": true, 00:09:11.110 "data_offset": 2048, 00:09:11.110 "data_size": 63488 00:09:11.110 }, 00:09:11.110 { 00:09:11.110 "name": "BaseBdev3", 00:09:11.110 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:11.110 "is_configured": true, 00:09:11.110 "data_offset": 2048, 00:09:11.110 "data_size": 63488 00:09:11.110 } 00:09:11.110 ] 00:09:11.110 }' 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.110 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.680 [2024-10-29 10:58:16.919534] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.680 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.681 "name": "Existed_Raid", 00:09:11.681 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:11.681 "strip_size_kb": 64, 00:09:11.681 "state": "configuring", 00:09:11.681 "raid_level": "concat", 00:09:11.681 "superblock": true, 00:09:11.681 "num_base_bdevs": 3, 00:09:11.681 "num_base_bdevs_discovered": 1, 00:09:11.681 "num_base_bdevs_operational": 3, 00:09:11.681 "base_bdevs_list": [ 00:09:11.681 { 00:09:11.681 "name": "BaseBdev1", 00:09:11.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.681 "is_configured": false, 00:09:11.681 "data_offset": 0, 00:09:11.681 "data_size": 0 00:09:11.681 }, 00:09:11.681 { 00:09:11.681 "name": null, 00:09:11.681 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:11.681 "is_configured": false, 00:09:11.681 "data_offset": 0, 00:09:11.681 "data_size": 63488 00:09:11.681 }, 00:09:11.681 { 00:09:11.681 "name": "BaseBdev3", 00:09:11.681 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:11.681 "is_configured": true, 00:09:11.681 "data_offset": 2048, 00:09:11.681 "data_size": 63488 00:09:11.681 } 00:09:11.681 ] 00:09:11.681 }' 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.681 10:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.941 [2024-10-29 10:58:17.396048] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:11.941 BaseBdev1 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.941 [ 00:09:11.941 { 00:09:11.941 "name": "BaseBdev1", 00:09:11.941 "aliases": [ 00:09:11.941 "bed5554a-5e4d-4c75-be51-671328982ec4" 00:09:11.941 ], 00:09:11.941 "product_name": "Malloc disk", 00:09:11.941 "block_size": 512, 00:09:11.941 "num_blocks": 65536, 00:09:11.941 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:11.941 "assigned_rate_limits": { 00:09:11.941 "rw_ios_per_sec": 0, 00:09:11.941 "rw_mbytes_per_sec": 0, 00:09:11.941 "r_mbytes_per_sec": 0, 00:09:11.941 "w_mbytes_per_sec": 0 00:09:11.941 }, 00:09:11.941 "claimed": true, 00:09:11.941 "claim_type": "exclusive_write", 00:09:11.941 "zoned": false, 00:09:11.941 "supported_io_types": { 00:09:11.941 "read": true, 00:09:11.941 "write": true, 00:09:11.941 "unmap": true, 00:09:11.941 "flush": true, 00:09:11.941 "reset": true, 00:09:11.941 "nvme_admin": false, 00:09:11.941 "nvme_io": false, 00:09:11.941 "nvme_io_md": false, 00:09:11.941 "write_zeroes": true, 00:09:11.941 "zcopy": true, 00:09:11.941 "get_zone_info": false, 00:09:11.941 "zone_management": false, 00:09:11.941 "zone_append": false, 00:09:11.941 "compare": false, 00:09:11.941 "compare_and_write": false, 00:09:11.941 "abort": true, 00:09:11.941 "seek_hole": false, 00:09:11.941 "seek_data": false, 00:09:11.941 "copy": true, 00:09:11.941 "nvme_iov_md": false 00:09:11.941 }, 00:09:11.941 "memory_domains": [ 00:09:11.941 { 00:09:11.941 "dma_device_id": "system", 00:09:11.941 "dma_device_type": 1 00:09:11.941 }, 00:09:11.941 { 00:09:11.941 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.941 "dma_device_type": 2 00:09:11.941 } 00:09:11.941 ], 00:09:11.941 "driver_specific": {} 00:09:11.941 } 00:09:11.941 ] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.941 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.201 "name": "Existed_Raid", 00:09:12.201 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:12.201 "strip_size_kb": 64, 00:09:12.201 "state": "configuring", 00:09:12.201 "raid_level": "concat", 00:09:12.201 "superblock": true, 00:09:12.201 "num_base_bdevs": 3, 00:09:12.201 "num_base_bdevs_discovered": 2, 00:09:12.201 "num_base_bdevs_operational": 3, 00:09:12.201 "base_bdevs_list": [ 00:09:12.201 { 00:09:12.201 "name": "BaseBdev1", 00:09:12.201 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:12.201 "is_configured": true, 00:09:12.201 "data_offset": 2048, 00:09:12.201 "data_size": 63488 00:09:12.201 }, 00:09:12.201 { 00:09:12.201 "name": null, 00:09:12.201 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:12.201 "is_configured": false, 00:09:12.201 "data_offset": 0, 00:09:12.201 "data_size": 63488 00:09:12.201 }, 00:09:12.201 { 00:09:12.201 "name": "BaseBdev3", 00:09:12.201 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:12.201 "is_configured": true, 00:09:12.201 "data_offset": 2048, 00:09:12.201 "data_size": 63488 00:09:12.201 } 00:09:12.201 ] 00:09:12.201 }' 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.201 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.462 [2024-10-29 10:58:17.875450] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.462 "name": "Existed_Raid", 00:09:12.462 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:12.462 "strip_size_kb": 64, 00:09:12.462 "state": "configuring", 00:09:12.462 "raid_level": "concat", 00:09:12.462 "superblock": true, 00:09:12.462 "num_base_bdevs": 3, 00:09:12.462 "num_base_bdevs_discovered": 1, 00:09:12.462 "num_base_bdevs_operational": 3, 00:09:12.462 "base_bdevs_list": [ 00:09:12.462 { 00:09:12.462 "name": "BaseBdev1", 00:09:12.462 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:12.462 "is_configured": true, 00:09:12.462 "data_offset": 2048, 00:09:12.462 "data_size": 63488 00:09:12.462 }, 00:09:12.462 { 00:09:12.462 "name": null, 00:09:12.462 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:12.462 "is_configured": false, 00:09:12.462 "data_offset": 0, 00:09:12.462 "data_size": 63488 00:09:12.462 }, 00:09:12.462 { 00:09:12.462 "name": null, 00:09:12.462 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:12.462 "is_configured": false, 00:09:12.462 "data_offset": 0, 00:09:12.462 "data_size": 63488 00:09:12.462 } 00:09:12.462 ] 00:09:12.462 }' 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.462 10:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.032 [2024-10-29 10:58:18.362677] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.032 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.032 "name": "Existed_Raid", 00:09:13.032 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:13.032 "strip_size_kb": 64, 00:09:13.032 "state": "configuring", 00:09:13.032 "raid_level": "concat", 00:09:13.032 "superblock": true, 00:09:13.032 "num_base_bdevs": 3, 00:09:13.032 "num_base_bdevs_discovered": 2, 00:09:13.032 "num_base_bdevs_operational": 3, 00:09:13.033 "base_bdevs_list": [ 00:09:13.033 { 00:09:13.033 "name": "BaseBdev1", 00:09:13.033 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:13.033 "is_configured": true, 00:09:13.033 "data_offset": 2048, 00:09:13.033 "data_size": 63488 00:09:13.033 }, 00:09:13.033 { 00:09:13.033 "name": null, 00:09:13.033 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:13.033 "is_configured": false, 00:09:13.033 "data_offset": 0, 00:09:13.033 "data_size": 63488 00:09:13.033 }, 00:09:13.033 { 00:09:13.033 "name": "BaseBdev3", 00:09:13.033 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:13.033 "is_configured": true, 00:09:13.033 "data_offset": 2048, 00:09:13.033 "data_size": 63488 00:09:13.033 } 00:09:13.033 ] 00:09:13.033 }' 00:09:13.033 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.033 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.602 [2024-10-29 10:58:18.857917] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.602 "name": "Existed_Raid", 00:09:13.602 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:13.602 "strip_size_kb": 64, 00:09:13.602 "state": "configuring", 00:09:13.602 "raid_level": "concat", 00:09:13.602 "superblock": true, 00:09:13.602 "num_base_bdevs": 3, 00:09:13.602 "num_base_bdevs_discovered": 1, 00:09:13.602 "num_base_bdevs_operational": 3, 00:09:13.602 "base_bdevs_list": [ 00:09:13.602 { 00:09:13.602 "name": null, 00:09:13.602 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:13.602 "is_configured": false, 00:09:13.602 "data_offset": 0, 00:09:13.602 "data_size": 63488 00:09:13.602 }, 00:09:13.602 { 00:09:13.602 "name": null, 00:09:13.602 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:13.602 "is_configured": false, 00:09:13.602 "data_offset": 0, 00:09:13.602 "data_size": 63488 00:09:13.602 }, 00:09:13.602 { 00:09:13.602 "name": "BaseBdev3", 00:09:13.602 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:13.602 "is_configured": true, 00:09:13.602 "data_offset": 2048, 00:09:13.602 "data_size": 63488 00:09:13.602 } 00:09:13.602 ] 00:09:13.602 }' 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.602 10:58:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.866 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.866 [2024-10-29 10:58:19.361340] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:14.126 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.127 "name": "Existed_Raid", 00:09:14.127 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:14.127 "strip_size_kb": 64, 00:09:14.127 "state": "configuring", 00:09:14.127 "raid_level": "concat", 00:09:14.127 "superblock": true, 00:09:14.127 "num_base_bdevs": 3, 00:09:14.127 "num_base_bdevs_discovered": 2, 00:09:14.127 "num_base_bdevs_operational": 3, 00:09:14.127 "base_bdevs_list": [ 00:09:14.127 { 00:09:14.127 "name": null, 00:09:14.127 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:14.127 "is_configured": false, 00:09:14.127 "data_offset": 0, 00:09:14.127 "data_size": 63488 00:09:14.127 }, 00:09:14.127 { 00:09:14.127 "name": "BaseBdev2", 00:09:14.127 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:14.127 "is_configured": true, 00:09:14.127 "data_offset": 2048, 00:09:14.127 "data_size": 63488 00:09:14.127 }, 00:09:14.127 { 00:09:14.127 "name": "BaseBdev3", 00:09:14.127 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:14.127 "is_configured": true, 00:09:14.127 "data_offset": 2048, 00:09:14.127 "data_size": 63488 00:09:14.127 } 00:09:14.127 ] 00:09:14.127 }' 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.127 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u bed5554a-5e4d-4c75-be51-671328982ec4 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.386 [2024-10-29 10:58:19.877235] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:14.386 [2024-10-29 10:58:19.877523] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:14.386 [2024-10-29 10:58:19.877578] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:14.386 [2024-10-29 10:58:19.877917] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:14.386 [2024-10-29 10:58:19.878084] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:14.386 [2024-10-29 10:58:19.878121] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:14.386 NewBaseBdev 00:09:14.386 [2024-10-29 10:58:19.878289] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.386 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.644 [ 00:09:14.644 { 00:09:14.644 "name": "NewBaseBdev", 00:09:14.644 "aliases": [ 00:09:14.644 "bed5554a-5e4d-4c75-be51-671328982ec4" 00:09:14.644 ], 00:09:14.644 "product_name": "Malloc disk", 00:09:14.644 "block_size": 512, 00:09:14.644 "num_blocks": 65536, 00:09:14.644 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:14.644 "assigned_rate_limits": { 00:09:14.644 "rw_ios_per_sec": 0, 00:09:14.644 "rw_mbytes_per_sec": 0, 00:09:14.644 "r_mbytes_per_sec": 0, 00:09:14.644 "w_mbytes_per_sec": 0 00:09:14.644 }, 00:09:14.644 "claimed": true, 00:09:14.644 "claim_type": "exclusive_write", 00:09:14.644 "zoned": false, 00:09:14.644 "supported_io_types": { 00:09:14.644 "read": true, 00:09:14.644 "write": true, 00:09:14.644 "unmap": true, 00:09:14.644 "flush": true, 00:09:14.644 "reset": true, 00:09:14.644 "nvme_admin": false, 00:09:14.644 "nvme_io": false, 00:09:14.644 "nvme_io_md": false, 00:09:14.644 "write_zeroes": true, 00:09:14.644 "zcopy": true, 00:09:14.644 "get_zone_info": false, 00:09:14.644 "zone_management": false, 00:09:14.644 "zone_append": false, 00:09:14.644 "compare": false, 00:09:14.644 "compare_and_write": false, 00:09:14.644 "abort": true, 00:09:14.644 "seek_hole": false, 00:09:14.644 "seek_data": false, 00:09:14.644 "copy": true, 00:09:14.644 "nvme_iov_md": false 00:09:14.644 }, 00:09:14.644 "memory_domains": [ 00:09:14.644 { 00:09:14.644 "dma_device_id": "system", 00:09:14.644 "dma_device_type": 1 00:09:14.644 }, 00:09:14.644 { 00:09:14.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.644 "dma_device_type": 2 00:09:14.644 } 00:09:14.644 ], 00:09:14.644 "driver_specific": {} 00:09:14.644 } 00:09:14.644 ] 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.644 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.644 "name": "Existed_Raid", 00:09:14.644 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:14.644 "strip_size_kb": 64, 00:09:14.644 "state": "online", 00:09:14.644 "raid_level": "concat", 00:09:14.644 "superblock": true, 00:09:14.644 "num_base_bdevs": 3, 00:09:14.644 "num_base_bdevs_discovered": 3, 00:09:14.645 "num_base_bdevs_operational": 3, 00:09:14.645 "base_bdevs_list": [ 00:09:14.645 { 00:09:14.645 "name": "NewBaseBdev", 00:09:14.645 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:14.645 "is_configured": true, 00:09:14.645 "data_offset": 2048, 00:09:14.645 "data_size": 63488 00:09:14.645 }, 00:09:14.645 { 00:09:14.645 "name": "BaseBdev2", 00:09:14.645 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:14.645 "is_configured": true, 00:09:14.645 "data_offset": 2048, 00:09:14.645 "data_size": 63488 00:09:14.645 }, 00:09:14.645 { 00:09:14.645 "name": "BaseBdev3", 00:09:14.645 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:14.645 "is_configured": true, 00:09:14.645 "data_offset": 2048, 00:09:14.645 "data_size": 63488 00:09:14.645 } 00:09:14.645 ] 00:09:14.645 }' 00:09:14.645 10:58:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.645 10:58:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:14.905 [2024-10-29 10:58:20.329013] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.905 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:14.905 "name": "Existed_Raid", 00:09:14.905 "aliases": [ 00:09:14.905 "995da598-9fa0-4488-956c-0ad814d7016a" 00:09:14.905 ], 00:09:14.905 "product_name": "Raid Volume", 00:09:14.905 "block_size": 512, 00:09:14.905 "num_blocks": 190464, 00:09:14.905 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:14.905 "assigned_rate_limits": { 00:09:14.905 "rw_ios_per_sec": 0, 00:09:14.905 "rw_mbytes_per_sec": 0, 00:09:14.905 "r_mbytes_per_sec": 0, 00:09:14.905 "w_mbytes_per_sec": 0 00:09:14.905 }, 00:09:14.905 "claimed": false, 00:09:14.905 "zoned": false, 00:09:14.906 "supported_io_types": { 00:09:14.906 "read": true, 00:09:14.906 "write": true, 00:09:14.906 "unmap": true, 00:09:14.906 "flush": true, 00:09:14.906 "reset": true, 00:09:14.906 "nvme_admin": false, 00:09:14.906 "nvme_io": false, 00:09:14.906 "nvme_io_md": false, 00:09:14.906 "write_zeroes": true, 00:09:14.906 "zcopy": false, 00:09:14.906 "get_zone_info": false, 00:09:14.906 "zone_management": false, 00:09:14.906 "zone_append": false, 00:09:14.906 "compare": false, 00:09:14.906 "compare_and_write": false, 00:09:14.906 "abort": false, 00:09:14.906 "seek_hole": false, 00:09:14.906 "seek_data": false, 00:09:14.906 "copy": false, 00:09:14.906 "nvme_iov_md": false 00:09:14.906 }, 00:09:14.906 "memory_domains": [ 00:09:14.906 { 00:09:14.906 "dma_device_id": "system", 00:09:14.906 "dma_device_type": 1 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.906 "dma_device_type": 2 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "dma_device_id": "system", 00:09:14.906 "dma_device_type": 1 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.906 "dma_device_type": 2 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "dma_device_id": "system", 00:09:14.906 "dma_device_type": 1 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.906 "dma_device_type": 2 00:09:14.906 } 00:09:14.906 ], 00:09:14.906 "driver_specific": { 00:09:14.906 "raid": { 00:09:14.906 "uuid": "995da598-9fa0-4488-956c-0ad814d7016a", 00:09:14.906 "strip_size_kb": 64, 00:09:14.906 "state": "online", 00:09:14.906 "raid_level": "concat", 00:09:14.906 "superblock": true, 00:09:14.906 "num_base_bdevs": 3, 00:09:14.906 "num_base_bdevs_discovered": 3, 00:09:14.906 "num_base_bdevs_operational": 3, 00:09:14.906 "base_bdevs_list": [ 00:09:14.906 { 00:09:14.906 "name": "NewBaseBdev", 00:09:14.906 "uuid": "bed5554a-5e4d-4c75-be51-671328982ec4", 00:09:14.906 "is_configured": true, 00:09:14.906 "data_offset": 2048, 00:09:14.906 "data_size": 63488 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "name": "BaseBdev2", 00:09:14.906 "uuid": "aa88c035-5474-4dae-a4ee-6e15e42a03be", 00:09:14.906 "is_configured": true, 00:09:14.906 "data_offset": 2048, 00:09:14.906 "data_size": 63488 00:09:14.906 }, 00:09:14.906 { 00:09:14.906 "name": "BaseBdev3", 00:09:14.906 "uuid": "3bbb7d66-56c9-4adc-a3f0-07c1df99242c", 00:09:14.906 "is_configured": true, 00:09:14.906 "data_offset": 2048, 00:09:14.906 "data_size": 63488 00:09:14.906 } 00:09:14.906 ] 00:09:14.906 } 00:09:14.906 } 00:09:14.906 }' 00:09:14.906 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:14.906 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:14.906 BaseBdev2 00:09:14.906 BaseBdev3' 00:09:14.906 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.169 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.170 [2024-10-29 10:58:20.556300] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:15.170 [2024-10-29 10:58:20.556352] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:15.170 [2024-10-29 10:58:20.556495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:15.170 [2024-10-29 10:58:20.556669] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:15.170 [2024-10-29 10:58:20.556689] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77492 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 77492 ']' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 77492 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 77492 00:09:15.170 killing process with pid 77492 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 77492' 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 77492 00:09:15.170 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 77492 00:09:15.170 [2024-10-29 10:58:20.593426] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:15.170 [2024-10-29 10:58:20.654034] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:15.741 10:58:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:15.741 00:09:15.741 real 0m8.677s 00:09:15.741 user 0m14.492s 00:09:15.741 sys 0m1.759s 00:09:15.741 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:15.741 ************************************ 00:09:15.741 END TEST raid_state_function_test_sb 00:09:15.741 ************************************ 00:09:15.741 10:58:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.741 10:58:21 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:09:15.741 10:58:21 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:09:15.741 10:58:21 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:15.741 10:58:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:15.741 ************************************ 00:09:15.741 START TEST raid_superblock_test 00:09:15.741 ************************************ 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test concat 3 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=78096 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 78096 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 78096 ']' 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:15.741 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:15.741 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.741 [2024-10-29 10:58:21.138663] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:15.741 [2024-10-29 10:58:21.139326] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78096 ] 00:09:16.001 [2024-10-29 10:58:21.307561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:16.001 [2024-10-29 10:58:21.346620] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:16.001 [2024-10-29 10:58:21.424041] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.001 [2024-10-29 10:58:21.424194] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.571 malloc1 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.571 [2024-10-29 10:58:21.983559] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:16.571 [2024-10-29 10:58:21.983707] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:16.571 [2024-10-29 10:58:21.983748] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:16.571 [2024-10-29 10:58:21.983816] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:16.571 [2024-10-29 10:58:21.986294] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:16.571 [2024-10-29 10:58:21.986334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:16.571 pt1 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.571 10:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.571 malloc2 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.571 [2024-10-29 10:58:22.022135] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:16.571 [2024-10-29 10:58:22.022242] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:16.571 [2024-10-29 10:58:22.022275] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:16.571 [2024-10-29 10:58:22.022306] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:16.571 [2024-10-29 10:58:22.024691] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:16.571 [2024-10-29 10:58:22.024761] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:16.571 pt2 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.571 malloc3 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:16.571 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.572 [2024-10-29 10:58:22.060665] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:16.572 [2024-10-29 10:58:22.060757] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:16.572 [2024-10-29 10:58:22.060793] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:16.572 [2024-10-29 10:58:22.060825] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:16.572 [2024-10-29 10:58:22.063207] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:16.572 [2024-10-29 10:58:22.063288] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:16.572 pt3 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.572 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.831 [2024-10-29 10:58:22.072679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:16.831 [2024-10-29 10:58:22.074881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:16.831 [2024-10-29 10:58:22.074982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:16.831 [2024-10-29 10:58:22.075204] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:09:16.831 [2024-10-29 10:58:22.075273] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:16.831 [2024-10-29 10:58:22.075633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:16.831 [2024-10-29 10:58:22.075867] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:09:16.831 [2024-10-29 10:58:22.075917] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:09:16.832 [2024-10-29 10:58:22.076130] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.832 "name": "raid_bdev1", 00:09:16.832 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:16.832 "strip_size_kb": 64, 00:09:16.832 "state": "online", 00:09:16.832 "raid_level": "concat", 00:09:16.832 "superblock": true, 00:09:16.832 "num_base_bdevs": 3, 00:09:16.832 "num_base_bdevs_discovered": 3, 00:09:16.832 "num_base_bdevs_operational": 3, 00:09:16.832 "base_bdevs_list": [ 00:09:16.832 { 00:09:16.832 "name": "pt1", 00:09:16.832 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:16.832 "is_configured": true, 00:09:16.832 "data_offset": 2048, 00:09:16.832 "data_size": 63488 00:09:16.832 }, 00:09:16.832 { 00:09:16.832 "name": "pt2", 00:09:16.832 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:16.832 "is_configured": true, 00:09:16.832 "data_offset": 2048, 00:09:16.832 "data_size": 63488 00:09:16.832 }, 00:09:16.832 { 00:09:16.832 "name": "pt3", 00:09:16.832 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:16.832 "is_configured": true, 00:09:16.832 "data_offset": 2048, 00:09:16.832 "data_size": 63488 00:09:16.832 } 00:09:16.832 ] 00:09:16.832 }' 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.832 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:17.092 [2024-10-29 10:58:22.528271] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:17.092 "name": "raid_bdev1", 00:09:17.092 "aliases": [ 00:09:17.092 "9e384914-2a07-4723-9a2b-2dd5a5a93d69" 00:09:17.092 ], 00:09:17.092 "product_name": "Raid Volume", 00:09:17.092 "block_size": 512, 00:09:17.092 "num_blocks": 190464, 00:09:17.092 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:17.092 "assigned_rate_limits": { 00:09:17.092 "rw_ios_per_sec": 0, 00:09:17.092 "rw_mbytes_per_sec": 0, 00:09:17.092 "r_mbytes_per_sec": 0, 00:09:17.092 "w_mbytes_per_sec": 0 00:09:17.092 }, 00:09:17.092 "claimed": false, 00:09:17.092 "zoned": false, 00:09:17.092 "supported_io_types": { 00:09:17.092 "read": true, 00:09:17.092 "write": true, 00:09:17.092 "unmap": true, 00:09:17.092 "flush": true, 00:09:17.092 "reset": true, 00:09:17.092 "nvme_admin": false, 00:09:17.092 "nvme_io": false, 00:09:17.092 "nvme_io_md": false, 00:09:17.092 "write_zeroes": true, 00:09:17.092 "zcopy": false, 00:09:17.092 "get_zone_info": false, 00:09:17.092 "zone_management": false, 00:09:17.092 "zone_append": false, 00:09:17.092 "compare": false, 00:09:17.092 "compare_and_write": false, 00:09:17.092 "abort": false, 00:09:17.092 "seek_hole": false, 00:09:17.092 "seek_data": false, 00:09:17.092 "copy": false, 00:09:17.092 "nvme_iov_md": false 00:09:17.092 }, 00:09:17.092 "memory_domains": [ 00:09:17.092 { 00:09:17.092 "dma_device_id": "system", 00:09:17.092 "dma_device_type": 1 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.092 "dma_device_type": 2 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "dma_device_id": "system", 00:09:17.092 "dma_device_type": 1 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.092 "dma_device_type": 2 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "dma_device_id": "system", 00:09:17.092 "dma_device_type": 1 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.092 "dma_device_type": 2 00:09:17.092 } 00:09:17.092 ], 00:09:17.092 "driver_specific": { 00:09:17.092 "raid": { 00:09:17.092 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:17.092 "strip_size_kb": 64, 00:09:17.092 "state": "online", 00:09:17.092 "raid_level": "concat", 00:09:17.092 "superblock": true, 00:09:17.092 "num_base_bdevs": 3, 00:09:17.092 "num_base_bdevs_discovered": 3, 00:09:17.092 "num_base_bdevs_operational": 3, 00:09:17.092 "base_bdevs_list": [ 00:09:17.092 { 00:09:17.092 "name": "pt1", 00:09:17.092 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:17.092 "is_configured": true, 00:09:17.092 "data_offset": 2048, 00:09:17.092 "data_size": 63488 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "name": "pt2", 00:09:17.092 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:17.092 "is_configured": true, 00:09:17.092 "data_offset": 2048, 00:09:17.092 "data_size": 63488 00:09:17.092 }, 00:09:17.092 { 00:09:17.092 "name": "pt3", 00:09:17.092 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:17.092 "is_configured": true, 00:09:17.092 "data_offset": 2048, 00:09:17.092 "data_size": 63488 00:09:17.092 } 00:09:17.092 ] 00:09:17.092 } 00:09:17.092 } 00:09:17.092 }' 00:09:17.092 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:17.352 pt2 00:09:17.352 pt3' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 [2024-10-29 10:58:22.783883] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=9e384914-2a07-4723-9a2b-2dd5a5a93d69 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 9e384914-2a07-4723-9a2b-2dd5a5a93d69 ']' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 [2024-10-29 10:58:22.819499] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:17.352 [2024-10-29 10:58:22.819547] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:17.352 [2024-10-29 10:58:22.819664] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:17.352 [2024-10-29 10:58:22.819734] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:17.352 [2024-10-29 10:58:22.819753] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.352 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.613 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.613 [2024-10-29 10:58:22.971297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:17.613 [2024-10-29 10:58:22.973715] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:17.613 [2024-10-29 10:58:22.973776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:17.613 [2024-10-29 10:58:22.973858] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:17.613 [2024-10-29 10:58:22.973913] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:17.613 [2024-10-29 10:58:22.973934] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:17.613 [2024-10-29 10:58:22.973948] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:17.613 [2024-10-29 10:58:22.973959] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:09:17.613 request: 00:09:17.613 { 00:09:17.613 "name": "raid_bdev1", 00:09:17.613 "raid_level": "concat", 00:09:17.613 "base_bdevs": [ 00:09:17.613 "malloc1", 00:09:17.613 "malloc2", 00:09:17.613 "malloc3" 00:09:17.613 ], 00:09:17.613 "strip_size_kb": 64, 00:09:17.613 "superblock": false, 00:09:17.613 "method": "bdev_raid_create", 00:09:17.613 "req_id": 1 00:09:17.613 } 00:09:17.613 Got JSON-RPC error response 00:09:17.613 response: 00:09:17.613 { 00:09:17.614 "code": -17, 00:09:17.614 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:17.614 } 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.614 10:58:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.614 [2024-10-29 10:58:23.023126] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:17.614 [2024-10-29 10:58:23.023303] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:17.614 [2024-10-29 10:58:23.023376] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:17.614 [2024-10-29 10:58:23.023417] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:17.614 [2024-10-29 10:58:23.026032] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:17.614 [2024-10-29 10:58:23.026122] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:17.614 [2024-10-29 10:58:23.026272] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:17.614 [2024-10-29 10:58:23.026389] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:17.614 pt1 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.614 "name": "raid_bdev1", 00:09:17.614 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:17.614 "strip_size_kb": 64, 00:09:17.614 "state": "configuring", 00:09:17.614 "raid_level": "concat", 00:09:17.614 "superblock": true, 00:09:17.614 "num_base_bdevs": 3, 00:09:17.614 "num_base_bdevs_discovered": 1, 00:09:17.614 "num_base_bdevs_operational": 3, 00:09:17.614 "base_bdevs_list": [ 00:09:17.614 { 00:09:17.614 "name": "pt1", 00:09:17.614 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:17.614 "is_configured": true, 00:09:17.614 "data_offset": 2048, 00:09:17.614 "data_size": 63488 00:09:17.614 }, 00:09:17.614 { 00:09:17.614 "name": null, 00:09:17.614 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:17.614 "is_configured": false, 00:09:17.614 "data_offset": 2048, 00:09:17.614 "data_size": 63488 00:09:17.614 }, 00:09:17.614 { 00:09:17.614 "name": null, 00:09:17.614 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:17.614 "is_configured": false, 00:09:17.614 "data_offset": 2048, 00:09:17.614 "data_size": 63488 00:09:17.614 } 00:09:17.614 ] 00:09:17.614 }' 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.614 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.185 [2024-10-29 10:58:23.442449] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:18.185 [2024-10-29 10:58:23.442656] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.185 [2024-10-29 10:58:23.442687] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:18.185 [2024-10-29 10:58:23.442703] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.185 [2024-10-29 10:58:23.443273] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.185 [2024-10-29 10:58:23.443306] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:18.185 [2024-10-29 10:58:23.443424] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:18.185 [2024-10-29 10:58:23.443459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:18.185 pt2 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.185 [2024-10-29 10:58:23.450434] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.185 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.186 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.186 "name": "raid_bdev1", 00:09:18.186 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:18.186 "strip_size_kb": 64, 00:09:18.186 "state": "configuring", 00:09:18.186 "raid_level": "concat", 00:09:18.186 "superblock": true, 00:09:18.186 "num_base_bdevs": 3, 00:09:18.186 "num_base_bdevs_discovered": 1, 00:09:18.186 "num_base_bdevs_operational": 3, 00:09:18.186 "base_bdevs_list": [ 00:09:18.186 { 00:09:18.186 "name": "pt1", 00:09:18.186 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:18.186 "is_configured": true, 00:09:18.186 "data_offset": 2048, 00:09:18.186 "data_size": 63488 00:09:18.186 }, 00:09:18.186 { 00:09:18.186 "name": null, 00:09:18.186 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:18.186 "is_configured": false, 00:09:18.186 "data_offset": 0, 00:09:18.186 "data_size": 63488 00:09:18.186 }, 00:09:18.186 { 00:09:18.186 "name": null, 00:09:18.186 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:18.186 "is_configured": false, 00:09:18.186 "data_offset": 2048, 00:09:18.186 "data_size": 63488 00:09:18.186 } 00:09:18.186 ] 00:09:18.186 }' 00:09:18.186 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.186 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.446 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:18.446 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:18.446 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:18.446 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.446 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.447 [2024-10-29 10:58:23.889655] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:18.447 [2024-10-29 10:58:23.889761] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.447 [2024-10-29 10:58:23.889791] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:18.447 [2024-10-29 10:58:23.889802] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.447 [2024-10-29 10:58:23.890296] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.447 [2024-10-29 10:58:23.890316] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:18.447 [2024-10-29 10:58:23.890428] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:18.447 [2024-10-29 10:58:23.890455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:18.447 pt2 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.447 [2024-10-29 10:58:23.901566] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:18.447 [2024-10-29 10:58:23.901621] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.447 [2024-10-29 10:58:23.901643] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:18.447 [2024-10-29 10:58:23.901652] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.447 [2024-10-29 10:58:23.902058] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.447 [2024-10-29 10:58:23.902073] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:18.447 [2024-10-29 10:58:23.902141] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:18.447 [2024-10-29 10:58:23.902159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:18.447 [2024-10-29 10:58:23.902262] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:18.447 [2024-10-29 10:58:23.902270] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:18.447 [2024-10-29 10:58:23.902524] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:18.447 [2024-10-29 10:58:23.902651] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:18.447 [2024-10-29 10:58:23.902663] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:09:18.447 [2024-10-29 10:58:23.902764] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:18.447 pt3 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.447 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.708 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.708 "name": "raid_bdev1", 00:09:18.708 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:18.708 "strip_size_kb": 64, 00:09:18.708 "state": "online", 00:09:18.708 "raid_level": "concat", 00:09:18.708 "superblock": true, 00:09:18.708 "num_base_bdevs": 3, 00:09:18.708 "num_base_bdevs_discovered": 3, 00:09:18.708 "num_base_bdevs_operational": 3, 00:09:18.708 "base_bdevs_list": [ 00:09:18.708 { 00:09:18.708 "name": "pt1", 00:09:18.708 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:18.708 "is_configured": true, 00:09:18.708 "data_offset": 2048, 00:09:18.708 "data_size": 63488 00:09:18.708 }, 00:09:18.708 { 00:09:18.708 "name": "pt2", 00:09:18.708 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:18.708 "is_configured": true, 00:09:18.708 "data_offset": 2048, 00:09:18.708 "data_size": 63488 00:09:18.708 }, 00:09:18.708 { 00:09:18.708 "name": "pt3", 00:09:18.708 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:18.708 "is_configured": true, 00:09:18.708 "data_offset": 2048, 00:09:18.708 "data_size": 63488 00:09:18.708 } 00:09:18.708 ] 00:09:18.708 }' 00:09:18.708 10:58:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.708 10:58:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.968 [2024-10-29 10:58:24.341174] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.968 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:18.968 "name": "raid_bdev1", 00:09:18.968 "aliases": [ 00:09:18.968 "9e384914-2a07-4723-9a2b-2dd5a5a93d69" 00:09:18.968 ], 00:09:18.968 "product_name": "Raid Volume", 00:09:18.968 "block_size": 512, 00:09:18.968 "num_blocks": 190464, 00:09:18.968 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:18.968 "assigned_rate_limits": { 00:09:18.968 "rw_ios_per_sec": 0, 00:09:18.968 "rw_mbytes_per_sec": 0, 00:09:18.968 "r_mbytes_per_sec": 0, 00:09:18.968 "w_mbytes_per_sec": 0 00:09:18.968 }, 00:09:18.968 "claimed": false, 00:09:18.968 "zoned": false, 00:09:18.968 "supported_io_types": { 00:09:18.968 "read": true, 00:09:18.968 "write": true, 00:09:18.968 "unmap": true, 00:09:18.968 "flush": true, 00:09:18.968 "reset": true, 00:09:18.968 "nvme_admin": false, 00:09:18.968 "nvme_io": false, 00:09:18.968 "nvme_io_md": false, 00:09:18.968 "write_zeroes": true, 00:09:18.968 "zcopy": false, 00:09:18.968 "get_zone_info": false, 00:09:18.968 "zone_management": false, 00:09:18.968 "zone_append": false, 00:09:18.968 "compare": false, 00:09:18.968 "compare_and_write": false, 00:09:18.968 "abort": false, 00:09:18.968 "seek_hole": false, 00:09:18.968 "seek_data": false, 00:09:18.968 "copy": false, 00:09:18.968 "nvme_iov_md": false 00:09:18.968 }, 00:09:18.968 "memory_domains": [ 00:09:18.968 { 00:09:18.969 "dma_device_id": "system", 00:09:18.969 "dma_device_type": 1 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.969 "dma_device_type": 2 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "dma_device_id": "system", 00:09:18.969 "dma_device_type": 1 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.969 "dma_device_type": 2 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "dma_device_id": "system", 00:09:18.969 "dma_device_type": 1 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.969 "dma_device_type": 2 00:09:18.969 } 00:09:18.969 ], 00:09:18.969 "driver_specific": { 00:09:18.969 "raid": { 00:09:18.969 "uuid": "9e384914-2a07-4723-9a2b-2dd5a5a93d69", 00:09:18.969 "strip_size_kb": 64, 00:09:18.969 "state": "online", 00:09:18.969 "raid_level": "concat", 00:09:18.969 "superblock": true, 00:09:18.969 "num_base_bdevs": 3, 00:09:18.969 "num_base_bdevs_discovered": 3, 00:09:18.969 "num_base_bdevs_operational": 3, 00:09:18.969 "base_bdevs_list": [ 00:09:18.969 { 00:09:18.969 "name": "pt1", 00:09:18.969 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:18.969 "is_configured": true, 00:09:18.969 "data_offset": 2048, 00:09:18.969 "data_size": 63488 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "name": "pt2", 00:09:18.969 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:18.969 "is_configured": true, 00:09:18.969 "data_offset": 2048, 00:09:18.969 "data_size": 63488 00:09:18.969 }, 00:09:18.969 { 00:09:18.969 "name": "pt3", 00:09:18.969 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:18.969 "is_configured": true, 00:09:18.969 "data_offset": 2048, 00:09:18.969 "data_size": 63488 00:09:18.969 } 00:09:18.969 ] 00:09:18.969 } 00:09:18.969 } 00:09:18.969 }' 00:09:18.969 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:18.969 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:18.969 pt2 00:09:18.969 pt3' 00:09:18.969 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.969 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.229 [2024-10-29 10:58:24.628753] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 9e384914-2a07-4723-9a2b-2dd5a5a93d69 '!=' 9e384914-2a07-4723-9a2b-2dd5a5a93d69 ']' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 78096 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 78096 ']' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 78096 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 78096 00:09:19.229 killing process with pid 78096 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 78096' 00:09:19.229 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 78096 00:09:19.230 [2024-10-29 10:58:24.701193] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:19.230 [2024-10-29 10:58:24.701321] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:19.230 10:58:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 78096 00:09:19.230 [2024-10-29 10:58:24.701412] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:19.230 [2024-10-29 10:58:24.701424] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:09:19.489 [2024-10-29 10:58:24.764195] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:19.749 10:58:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:19.749 00:09:19.749 real 0m4.035s 00:09:19.749 user 0m6.193s 00:09:19.749 sys 0m0.887s 00:09:19.749 10:58:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:19.749 ************************************ 00:09:19.749 END TEST raid_superblock_test 00:09:19.749 ************************************ 00:09:19.749 10:58:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.749 10:58:25 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:09:19.749 10:58:25 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:19.749 10:58:25 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:19.749 10:58:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:19.749 ************************************ 00:09:19.749 START TEST raid_read_error_test 00:09:19.749 ************************************ 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 3 read 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.OOG0K2LwIa 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78338 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78338 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 78338 ']' 00:09:19.749 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:19.749 10:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.010 [2024-10-29 10:58:25.258115] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:20.010 [2024-10-29 10:58:25.258236] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78338 ] 00:09:20.010 [2024-10-29 10:58:25.431117] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:20.010 [2024-10-29 10:58:25.472580] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:20.269 [2024-10-29 10:58:25.550156] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:20.269 [2024-10-29 10:58:25.550299] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 BaseBdev1_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 true 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 [2024-10-29 10:58:26.129142] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:20.840 [2024-10-29 10:58:26.129284] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:20.840 [2024-10-29 10:58:26.129335] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:20.840 [2024-10-29 10:58:26.129413] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:20.840 [2024-10-29 10:58:26.132034] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:20.840 [2024-10-29 10:58:26.132115] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:20.840 BaseBdev1 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 BaseBdev2_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 true 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 [2024-10-29 10:58:26.176404] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:20.840 [2024-10-29 10:58:26.176482] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:20.840 [2024-10-29 10:58:26.176505] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:20.840 [2024-10-29 10:58:26.176517] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:20.840 [2024-10-29 10:58:26.179077] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:20.840 [2024-10-29 10:58:26.179118] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:20.840 BaseBdev2 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 BaseBdev3_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 true 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.840 [2024-10-29 10:58:26.223366] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:20.840 [2024-10-29 10:58:26.223439] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:20.840 [2024-10-29 10:58:26.223462] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:20.840 [2024-10-29 10:58:26.223472] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:20.840 [2024-10-29 10:58:26.225872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:20.840 [2024-10-29 10:58:26.225911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:20.840 BaseBdev3 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.840 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.841 [2024-10-29 10:58:26.235449] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:20.841 [2024-10-29 10:58:26.237664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:20.841 [2024-10-29 10:58:26.237756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:20.841 [2024-10-29 10:58:26.237955] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:20.841 [2024-10-29 10:58:26.237971] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:20.841 [2024-10-29 10:58:26.238269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:20.841 [2024-10-29 10:58:26.238452] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:20.841 [2024-10-29 10:58:26.238464] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:20.841 [2024-10-29 10:58:26.238632] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.841 "name": "raid_bdev1", 00:09:20.841 "uuid": "752c6d2b-6d5a-4808-9f3a-a6cc640ec865", 00:09:20.841 "strip_size_kb": 64, 00:09:20.841 "state": "online", 00:09:20.841 "raid_level": "concat", 00:09:20.841 "superblock": true, 00:09:20.841 "num_base_bdevs": 3, 00:09:20.841 "num_base_bdevs_discovered": 3, 00:09:20.841 "num_base_bdevs_operational": 3, 00:09:20.841 "base_bdevs_list": [ 00:09:20.841 { 00:09:20.841 "name": "BaseBdev1", 00:09:20.841 "uuid": "ba4eace8-484c-5cfc-a7fd-f78455add4c5", 00:09:20.841 "is_configured": true, 00:09:20.841 "data_offset": 2048, 00:09:20.841 "data_size": 63488 00:09:20.841 }, 00:09:20.841 { 00:09:20.841 "name": "BaseBdev2", 00:09:20.841 "uuid": "794af186-2aac-5f1a-8211-a886ab3eb867", 00:09:20.841 "is_configured": true, 00:09:20.841 "data_offset": 2048, 00:09:20.841 "data_size": 63488 00:09:20.841 }, 00:09:20.841 { 00:09:20.841 "name": "BaseBdev3", 00:09:20.841 "uuid": "a7cad73b-3bef-5f94-91d7-7b4ab6b1daa4", 00:09:20.841 "is_configured": true, 00:09:20.841 "data_offset": 2048, 00:09:20.841 "data_size": 63488 00:09:20.841 } 00:09:20.841 ] 00:09:20.841 }' 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.841 10:58:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.412 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:21.412 10:58:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:21.412 [2024-10-29 10:58:26.782945] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.350 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.351 "name": "raid_bdev1", 00:09:22.351 "uuid": "752c6d2b-6d5a-4808-9f3a-a6cc640ec865", 00:09:22.351 "strip_size_kb": 64, 00:09:22.351 "state": "online", 00:09:22.351 "raid_level": "concat", 00:09:22.351 "superblock": true, 00:09:22.351 "num_base_bdevs": 3, 00:09:22.351 "num_base_bdevs_discovered": 3, 00:09:22.351 "num_base_bdevs_operational": 3, 00:09:22.351 "base_bdevs_list": [ 00:09:22.351 { 00:09:22.351 "name": "BaseBdev1", 00:09:22.351 "uuid": "ba4eace8-484c-5cfc-a7fd-f78455add4c5", 00:09:22.351 "is_configured": true, 00:09:22.351 "data_offset": 2048, 00:09:22.351 "data_size": 63488 00:09:22.351 }, 00:09:22.351 { 00:09:22.351 "name": "BaseBdev2", 00:09:22.351 "uuid": "794af186-2aac-5f1a-8211-a886ab3eb867", 00:09:22.351 "is_configured": true, 00:09:22.351 "data_offset": 2048, 00:09:22.351 "data_size": 63488 00:09:22.351 }, 00:09:22.351 { 00:09:22.351 "name": "BaseBdev3", 00:09:22.351 "uuid": "a7cad73b-3bef-5f94-91d7-7b4ab6b1daa4", 00:09:22.351 "is_configured": true, 00:09:22.351 "data_offset": 2048, 00:09:22.351 "data_size": 63488 00:09:22.351 } 00:09:22.351 ] 00:09:22.351 }' 00:09:22.351 10:58:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.351 10:58:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.920 [2024-10-29 10:58:28.140454] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:22.920 [2024-10-29 10:58:28.140506] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:22.920 [2024-10-29 10:58:28.143267] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:22.920 [2024-10-29 10:58:28.143332] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.920 [2024-10-29 10:58:28.143391] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:22.920 [2024-10-29 10:58:28.143407] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:22.920 { 00:09:22.920 "results": [ 00:09:22.920 { 00:09:22.920 "job": "raid_bdev1", 00:09:22.920 "core_mask": "0x1", 00:09:22.920 "workload": "randrw", 00:09:22.920 "percentage": 50, 00:09:22.920 "status": "finished", 00:09:22.920 "queue_depth": 1, 00:09:22.920 "io_size": 131072, 00:09:22.920 "runtime": 1.357903, 00:09:22.920 "iops": 14283.789048260443, 00:09:22.920 "mibps": 1785.4736310325554, 00:09:22.920 "io_failed": 1, 00:09:22.920 "io_timeout": 0, 00:09:22.920 "avg_latency_us": 98.42815597694057, 00:09:22.920 "min_latency_us": 24.034934497816593, 00:09:22.920 "max_latency_us": 1380.8349344978167 00:09:22.920 } 00:09:22.920 ], 00:09:22.920 "core_count": 1 00:09:22.920 } 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78338 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 78338 ']' 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 78338 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 78338 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:22.920 killing process with pid 78338 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 78338' 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 78338 00:09:22.920 [2024-10-29 10:58:28.175992] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:22.920 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 78338 00:09:22.920 [2024-10-29 10:58:28.224424] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.OOG0K2LwIa 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:09:23.180 ************************************ 00:09:23.180 END TEST raid_read_error_test 00:09:23.180 ************************************ 00:09:23.180 00:09:23.180 real 0m3.400s 00:09:23.180 user 0m4.194s 00:09:23.180 sys 0m0.582s 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:23.180 10:58:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.180 10:58:28 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:09:23.180 10:58:28 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:23.180 10:58:28 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:23.180 10:58:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:23.180 ************************************ 00:09:23.180 START TEST raid_write_error_test 00:09:23.180 ************************************ 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 3 write 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.0vfIvnTg3l 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78467 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:23.180 10:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78467 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 78467 ']' 00:09:23.181 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:23.181 10:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.441 [2024-10-29 10:58:28.726172] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:23.441 [2024-10-29 10:58:28.726312] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78467 ] 00:09:23.441 [2024-10-29 10:58:28.898632] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:23.441 [2024-10-29 10:58:28.937042] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:23.701 [2024-10-29 10:58:29.013793] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:23.701 [2024-10-29 10:58:29.013834] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 BaseBdev1_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 true 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 [2024-10-29 10:58:29.589151] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:24.273 [2024-10-29 10:58:29.589222] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.273 [2024-10-29 10:58:29.589245] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:24.273 [2024-10-29 10:58:29.589255] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:24.273 [2024-10-29 10:58:29.591738] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:24.273 [2024-10-29 10:58:29.591851] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:24.273 BaseBdev1 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 BaseBdev2_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 true 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 [2024-10-29 10:58:29.635970] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:24.273 [2024-10-29 10:58:29.636043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.273 [2024-10-29 10:58:29.636069] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:24.273 [2024-10-29 10:58:29.636079] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:24.273 [2024-10-29 10:58:29.638515] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:24.273 [2024-10-29 10:58:29.638627] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:24.273 BaseBdev2 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 BaseBdev3_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 true 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 [2024-10-29 10:58:29.682853] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:24.273 [2024-10-29 10:58:29.682914] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.273 [2024-10-29 10:58:29.682935] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:24.273 [2024-10-29 10:58:29.682945] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:24.273 [2024-10-29 10:58:29.685300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:24.273 [2024-10-29 10:58:29.685417] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:24.273 BaseBdev3 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.273 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.273 [2024-10-29 10:58:29.694908] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:24.273 [2024-10-29 10:58:29.697014] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:24.273 [2024-10-29 10:58:29.697154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:24.273 [2024-10-29 10:58:29.697366] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:24.273 [2024-10-29 10:58:29.697383] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:24.273 [2024-10-29 10:58:29.697662] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:24.273 [2024-10-29 10:58:29.697832] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:24.274 [2024-10-29 10:58:29.697842] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:24.274 [2024-10-29 10:58:29.697973] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.274 "name": "raid_bdev1", 00:09:24.274 "uuid": "f0f574ef-4527-4108-9b0c-9606ec83bc62", 00:09:24.274 "strip_size_kb": 64, 00:09:24.274 "state": "online", 00:09:24.274 "raid_level": "concat", 00:09:24.274 "superblock": true, 00:09:24.274 "num_base_bdevs": 3, 00:09:24.274 "num_base_bdevs_discovered": 3, 00:09:24.274 "num_base_bdevs_operational": 3, 00:09:24.274 "base_bdevs_list": [ 00:09:24.274 { 00:09:24.274 "name": "BaseBdev1", 00:09:24.274 "uuid": "adc7d110-dae6-596e-b4fe-c88a74303f15", 00:09:24.274 "is_configured": true, 00:09:24.274 "data_offset": 2048, 00:09:24.274 "data_size": 63488 00:09:24.274 }, 00:09:24.274 { 00:09:24.274 "name": "BaseBdev2", 00:09:24.274 "uuid": "df93e95e-47a6-5188-a182-09642e878e04", 00:09:24.274 "is_configured": true, 00:09:24.274 "data_offset": 2048, 00:09:24.274 "data_size": 63488 00:09:24.274 }, 00:09:24.274 { 00:09:24.274 "name": "BaseBdev3", 00:09:24.274 "uuid": "d7979861-b739-5953-bf06-c35d810520b7", 00:09:24.274 "is_configured": true, 00:09:24.274 "data_offset": 2048, 00:09:24.274 "data_size": 63488 00:09:24.274 } 00:09:24.274 ] 00:09:24.274 }' 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.274 10:58:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.859 10:58:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:24.859 10:58:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:24.859 [2024-10-29 10:58:30.266528] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.799 "name": "raid_bdev1", 00:09:25.799 "uuid": "f0f574ef-4527-4108-9b0c-9606ec83bc62", 00:09:25.799 "strip_size_kb": 64, 00:09:25.799 "state": "online", 00:09:25.799 "raid_level": "concat", 00:09:25.799 "superblock": true, 00:09:25.799 "num_base_bdevs": 3, 00:09:25.799 "num_base_bdevs_discovered": 3, 00:09:25.799 "num_base_bdevs_operational": 3, 00:09:25.799 "base_bdevs_list": [ 00:09:25.799 { 00:09:25.799 "name": "BaseBdev1", 00:09:25.799 "uuid": "adc7d110-dae6-596e-b4fe-c88a74303f15", 00:09:25.799 "is_configured": true, 00:09:25.799 "data_offset": 2048, 00:09:25.799 "data_size": 63488 00:09:25.799 }, 00:09:25.799 { 00:09:25.799 "name": "BaseBdev2", 00:09:25.799 "uuid": "df93e95e-47a6-5188-a182-09642e878e04", 00:09:25.799 "is_configured": true, 00:09:25.799 "data_offset": 2048, 00:09:25.799 "data_size": 63488 00:09:25.799 }, 00:09:25.799 { 00:09:25.799 "name": "BaseBdev3", 00:09:25.799 "uuid": "d7979861-b739-5953-bf06-c35d810520b7", 00:09:25.799 "is_configured": true, 00:09:25.799 "data_offset": 2048, 00:09:25.799 "data_size": 63488 00:09:25.799 } 00:09:25.799 ] 00:09:25.799 }' 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.799 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.369 [2024-10-29 10:58:31.619913] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:26.369 [2024-10-29 10:58:31.620052] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:26.369 [2024-10-29 10:58:31.622703] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:26.369 [2024-10-29 10:58:31.622806] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:26.369 [2024-10-29 10:58:31.622868] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:26.369 [2024-10-29 10:58:31.622935] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:26.369 { 00:09:26.369 "results": [ 00:09:26.369 { 00:09:26.369 "job": "raid_bdev1", 00:09:26.369 "core_mask": "0x1", 00:09:26.369 "workload": "randrw", 00:09:26.369 "percentage": 50, 00:09:26.369 "status": "finished", 00:09:26.369 "queue_depth": 1, 00:09:26.369 "io_size": 131072, 00:09:26.369 "runtime": 1.35367, 00:09:26.369 "iops": 14025.574918554743, 00:09:26.369 "mibps": 1753.1968648193429, 00:09:26.369 "io_failed": 1, 00:09:26.369 "io_timeout": 0, 00:09:26.369 "avg_latency_us": 100.1088651094992, 00:09:26.369 "min_latency_us": 26.270742358078603, 00:09:26.369 "max_latency_us": 1395.1441048034935 00:09:26.369 } 00:09:26.369 ], 00:09:26.369 "core_count": 1 00:09:26.369 } 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78467 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 78467 ']' 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 78467 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 78467 00:09:26.369 killing process with pid 78467 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 78467' 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 78467 00:09:26.369 [2024-10-29 10:58:31.656033] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:26.369 10:58:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 78467 00:09:26.369 [2024-10-29 10:58:31.705823] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.0vfIvnTg3l 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:26.628 ************************************ 00:09:26.628 END TEST raid_write_error_test 00:09:26.628 ************************************ 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:09:26.628 00:09:26.628 real 0m3.413s 00:09:26.628 user 0m4.220s 00:09:26.628 sys 0m0.593s 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:26.628 10:58:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.628 10:58:32 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:26.628 10:58:32 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:09:26.628 10:58:32 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:26.628 10:58:32 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:26.628 10:58:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:26.628 ************************************ 00:09:26.628 START TEST raid_state_function_test 00:09:26.628 ************************************ 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 3 false 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78594 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78594' 00:09:26.628 Process raid pid: 78594 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78594 00:09:26.628 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 78594 ']' 00:09:26.888 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:26.888 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:26.888 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:26.888 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:26.888 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:26.888 10:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.888 [2024-10-29 10:58:32.202347] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:26.888 [2024-10-29 10:58:32.202483] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:26.888 [2024-10-29 10:58:32.351282] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:27.147 [2024-10-29 10:58:32.395630] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:27.147 [2024-10-29 10:58:32.473234] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:27.147 [2024-10-29 10:58:32.473277] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.715 [2024-10-29 10:58:33.036667] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:27.715 [2024-10-29 10:58:33.036747] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:27.715 [2024-10-29 10:58:33.036763] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:27.715 [2024-10-29 10:58:33.036774] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:27.715 [2024-10-29 10:58:33.036781] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:27.715 [2024-10-29 10:58:33.036794] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.715 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.715 "name": "Existed_Raid", 00:09:27.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.715 "strip_size_kb": 0, 00:09:27.715 "state": "configuring", 00:09:27.715 "raid_level": "raid1", 00:09:27.715 "superblock": false, 00:09:27.715 "num_base_bdevs": 3, 00:09:27.715 "num_base_bdevs_discovered": 0, 00:09:27.715 "num_base_bdevs_operational": 3, 00:09:27.715 "base_bdevs_list": [ 00:09:27.715 { 00:09:27.715 "name": "BaseBdev1", 00:09:27.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.716 "is_configured": false, 00:09:27.716 "data_offset": 0, 00:09:27.716 "data_size": 0 00:09:27.716 }, 00:09:27.716 { 00:09:27.716 "name": "BaseBdev2", 00:09:27.716 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.716 "is_configured": false, 00:09:27.716 "data_offset": 0, 00:09:27.716 "data_size": 0 00:09:27.716 }, 00:09:27.716 { 00:09:27.716 "name": "BaseBdev3", 00:09:27.716 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.716 "is_configured": false, 00:09:27.716 "data_offset": 0, 00:09:27.716 "data_size": 0 00:09:27.716 } 00:09:27.716 ] 00:09:27.716 }' 00:09:27.716 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.716 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 [2024-10-29 10:58:33.487813] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:28.283 [2024-10-29 10:58:33.487993] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 [2024-10-29 10:58:33.499785] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:28.283 [2024-10-29 10:58:33.499894] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:28.283 [2024-10-29 10:58:33.499946] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:28.283 [2024-10-29 10:58:33.499991] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:28.283 [2024-10-29 10:58:33.500033] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:28.283 [2024-10-29 10:58:33.500078] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 [2024-10-29 10:58:33.527032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:28.283 BaseBdev1 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 [ 00:09:28.283 { 00:09:28.283 "name": "BaseBdev1", 00:09:28.283 "aliases": [ 00:09:28.283 "15c556fa-79b3-4f7c-99a0-41f7100b22c7" 00:09:28.283 ], 00:09:28.283 "product_name": "Malloc disk", 00:09:28.283 "block_size": 512, 00:09:28.283 "num_blocks": 65536, 00:09:28.283 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:28.283 "assigned_rate_limits": { 00:09:28.283 "rw_ios_per_sec": 0, 00:09:28.283 "rw_mbytes_per_sec": 0, 00:09:28.283 "r_mbytes_per_sec": 0, 00:09:28.283 "w_mbytes_per_sec": 0 00:09:28.283 }, 00:09:28.283 "claimed": true, 00:09:28.283 "claim_type": "exclusive_write", 00:09:28.283 "zoned": false, 00:09:28.283 "supported_io_types": { 00:09:28.283 "read": true, 00:09:28.283 "write": true, 00:09:28.283 "unmap": true, 00:09:28.283 "flush": true, 00:09:28.283 "reset": true, 00:09:28.283 "nvme_admin": false, 00:09:28.283 "nvme_io": false, 00:09:28.283 "nvme_io_md": false, 00:09:28.283 "write_zeroes": true, 00:09:28.283 "zcopy": true, 00:09:28.283 "get_zone_info": false, 00:09:28.283 "zone_management": false, 00:09:28.283 "zone_append": false, 00:09:28.283 "compare": false, 00:09:28.283 "compare_and_write": false, 00:09:28.283 "abort": true, 00:09:28.283 "seek_hole": false, 00:09:28.283 "seek_data": false, 00:09:28.283 "copy": true, 00:09:28.283 "nvme_iov_md": false 00:09:28.283 }, 00:09:28.283 "memory_domains": [ 00:09:28.283 { 00:09:28.283 "dma_device_id": "system", 00:09:28.283 "dma_device_type": 1 00:09:28.283 }, 00:09:28.283 { 00:09:28.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.283 "dma_device_type": 2 00:09:28.283 } 00:09:28.283 ], 00:09:28.283 "driver_specific": {} 00:09:28.283 } 00:09:28.283 ] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.283 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.283 "name": "Existed_Raid", 00:09:28.283 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.283 "strip_size_kb": 0, 00:09:28.283 "state": "configuring", 00:09:28.283 "raid_level": "raid1", 00:09:28.283 "superblock": false, 00:09:28.283 "num_base_bdevs": 3, 00:09:28.283 "num_base_bdevs_discovered": 1, 00:09:28.283 "num_base_bdevs_operational": 3, 00:09:28.283 "base_bdevs_list": [ 00:09:28.283 { 00:09:28.283 "name": "BaseBdev1", 00:09:28.284 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:28.284 "is_configured": true, 00:09:28.284 "data_offset": 0, 00:09:28.284 "data_size": 65536 00:09:28.284 }, 00:09:28.284 { 00:09:28.284 "name": "BaseBdev2", 00:09:28.284 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.284 "is_configured": false, 00:09:28.284 "data_offset": 0, 00:09:28.284 "data_size": 0 00:09:28.284 }, 00:09:28.284 { 00:09:28.284 "name": "BaseBdev3", 00:09:28.284 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.284 "is_configured": false, 00:09:28.284 "data_offset": 0, 00:09:28.284 "data_size": 0 00:09:28.284 } 00:09:28.284 ] 00:09:28.284 }' 00:09:28.284 10:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.284 10:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 [2024-10-29 10:58:34.014282] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:28.543 [2024-10-29 10:58:34.014365] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 [2024-10-29 10:58:34.026311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:28.543 [2024-10-29 10:58:34.028671] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:28.543 [2024-10-29 10:58:34.028727] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:28.543 [2024-10-29 10:58:34.028737] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:28.543 [2024-10-29 10:58:34.028747] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.543 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.802 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.802 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.802 "name": "Existed_Raid", 00:09:28.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.802 "strip_size_kb": 0, 00:09:28.802 "state": "configuring", 00:09:28.802 "raid_level": "raid1", 00:09:28.802 "superblock": false, 00:09:28.802 "num_base_bdevs": 3, 00:09:28.802 "num_base_bdevs_discovered": 1, 00:09:28.802 "num_base_bdevs_operational": 3, 00:09:28.802 "base_bdevs_list": [ 00:09:28.802 { 00:09:28.802 "name": "BaseBdev1", 00:09:28.802 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:28.802 "is_configured": true, 00:09:28.802 "data_offset": 0, 00:09:28.802 "data_size": 65536 00:09:28.802 }, 00:09:28.802 { 00:09:28.802 "name": "BaseBdev2", 00:09:28.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.802 "is_configured": false, 00:09:28.802 "data_offset": 0, 00:09:28.802 "data_size": 0 00:09:28.802 }, 00:09:28.802 { 00:09:28.802 "name": "BaseBdev3", 00:09:28.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.802 "is_configured": false, 00:09:28.802 "data_offset": 0, 00:09:28.802 "data_size": 0 00:09:28.802 } 00:09:28.802 ] 00:09:28.802 }' 00:09:28.802 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.802 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.062 [2024-10-29 10:58:34.438619] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:29.062 BaseBdev2 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:29.062 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.063 [ 00:09:29.063 { 00:09:29.063 "name": "BaseBdev2", 00:09:29.063 "aliases": [ 00:09:29.063 "2514c8d5-bc66-4d24-ac63-4858f5033ba9" 00:09:29.063 ], 00:09:29.063 "product_name": "Malloc disk", 00:09:29.063 "block_size": 512, 00:09:29.063 "num_blocks": 65536, 00:09:29.063 "uuid": "2514c8d5-bc66-4d24-ac63-4858f5033ba9", 00:09:29.063 "assigned_rate_limits": { 00:09:29.063 "rw_ios_per_sec": 0, 00:09:29.063 "rw_mbytes_per_sec": 0, 00:09:29.063 "r_mbytes_per_sec": 0, 00:09:29.063 "w_mbytes_per_sec": 0 00:09:29.063 }, 00:09:29.063 "claimed": true, 00:09:29.063 "claim_type": "exclusive_write", 00:09:29.063 "zoned": false, 00:09:29.063 "supported_io_types": { 00:09:29.063 "read": true, 00:09:29.063 "write": true, 00:09:29.063 "unmap": true, 00:09:29.063 "flush": true, 00:09:29.063 "reset": true, 00:09:29.063 "nvme_admin": false, 00:09:29.063 "nvme_io": false, 00:09:29.063 "nvme_io_md": false, 00:09:29.063 "write_zeroes": true, 00:09:29.063 "zcopy": true, 00:09:29.063 "get_zone_info": false, 00:09:29.063 "zone_management": false, 00:09:29.063 "zone_append": false, 00:09:29.063 "compare": false, 00:09:29.063 "compare_and_write": false, 00:09:29.063 "abort": true, 00:09:29.063 "seek_hole": false, 00:09:29.063 "seek_data": false, 00:09:29.063 "copy": true, 00:09:29.063 "nvme_iov_md": false 00:09:29.063 }, 00:09:29.063 "memory_domains": [ 00:09:29.063 { 00:09:29.063 "dma_device_id": "system", 00:09:29.063 "dma_device_type": 1 00:09:29.063 }, 00:09:29.063 { 00:09:29.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.063 "dma_device_type": 2 00:09:29.063 } 00:09:29.063 ], 00:09:29.063 "driver_specific": {} 00:09:29.063 } 00:09:29.063 ] 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.063 "name": "Existed_Raid", 00:09:29.063 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.063 "strip_size_kb": 0, 00:09:29.063 "state": "configuring", 00:09:29.063 "raid_level": "raid1", 00:09:29.063 "superblock": false, 00:09:29.063 "num_base_bdevs": 3, 00:09:29.063 "num_base_bdevs_discovered": 2, 00:09:29.063 "num_base_bdevs_operational": 3, 00:09:29.063 "base_bdevs_list": [ 00:09:29.063 { 00:09:29.063 "name": "BaseBdev1", 00:09:29.063 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:29.063 "is_configured": true, 00:09:29.063 "data_offset": 0, 00:09:29.063 "data_size": 65536 00:09:29.063 }, 00:09:29.063 { 00:09:29.063 "name": "BaseBdev2", 00:09:29.063 "uuid": "2514c8d5-bc66-4d24-ac63-4858f5033ba9", 00:09:29.063 "is_configured": true, 00:09:29.063 "data_offset": 0, 00:09:29.063 "data_size": 65536 00:09:29.063 }, 00:09:29.063 { 00:09:29.063 "name": "BaseBdev3", 00:09:29.063 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.063 "is_configured": false, 00:09:29.063 "data_offset": 0, 00:09:29.063 "data_size": 0 00:09:29.063 } 00:09:29.063 ] 00:09:29.063 }' 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.063 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.633 [2024-10-29 10:58:34.925743] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:29.633 [2024-10-29 10:58:34.925818] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:29.633 [2024-10-29 10:58:34.925834] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:29.633 [2024-10-29 10:58:34.926247] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:29.633 [2024-10-29 10:58:34.926480] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:29.633 [2024-10-29 10:58:34.926525] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:29.633 [2024-10-29 10:58:34.926785] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:29.633 BaseBdev3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.633 [ 00:09:29.633 { 00:09:29.633 "name": "BaseBdev3", 00:09:29.633 "aliases": [ 00:09:29.633 "b1cb77e0-4d4d-45e2-974d-a93e660ff1dc" 00:09:29.633 ], 00:09:29.633 "product_name": "Malloc disk", 00:09:29.633 "block_size": 512, 00:09:29.633 "num_blocks": 65536, 00:09:29.633 "uuid": "b1cb77e0-4d4d-45e2-974d-a93e660ff1dc", 00:09:29.633 "assigned_rate_limits": { 00:09:29.633 "rw_ios_per_sec": 0, 00:09:29.633 "rw_mbytes_per_sec": 0, 00:09:29.633 "r_mbytes_per_sec": 0, 00:09:29.633 "w_mbytes_per_sec": 0 00:09:29.633 }, 00:09:29.633 "claimed": true, 00:09:29.633 "claim_type": "exclusive_write", 00:09:29.633 "zoned": false, 00:09:29.633 "supported_io_types": { 00:09:29.633 "read": true, 00:09:29.633 "write": true, 00:09:29.633 "unmap": true, 00:09:29.633 "flush": true, 00:09:29.633 "reset": true, 00:09:29.633 "nvme_admin": false, 00:09:29.633 "nvme_io": false, 00:09:29.633 "nvme_io_md": false, 00:09:29.633 "write_zeroes": true, 00:09:29.633 "zcopy": true, 00:09:29.633 "get_zone_info": false, 00:09:29.633 "zone_management": false, 00:09:29.633 "zone_append": false, 00:09:29.633 "compare": false, 00:09:29.633 "compare_and_write": false, 00:09:29.633 "abort": true, 00:09:29.633 "seek_hole": false, 00:09:29.633 "seek_data": false, 00:09:29.633 "copy": true, 00:09:29.633 "nvme_iov_md": false 00:09:29.633 }, 00:09:29.633 "memory_domains": [ 00:09:29.633 { 00:09:29.633 "dma_device_id": "system", 00:09:29.633 "dma_device_type": 1 00:09:29.633 }, 00:09:29.633 { 00:09:29.633 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.633 "dma_device_type": 2 00:09:29.633 } 00:09:29.633 ], 00:09:29.633 "driver_specific": {} 00:09:29.633 } 00:09:29.633 ] 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.633 10:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.633 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.633 "name": "Existed_Raid", 00:09:29.633 "uuid": "0ceb04a2-b815-4dd8-9132-6f956ba17043", 00:09:29.633 "strip_size_kb": 0, 00:09:29.633 "state": "online", 00:09:29.633 "raid_level": "raid1", 00:09:29.633 "superblock": false, 00:09:29.633 "num_base_bdevs": 3, 00:09:29.633 "num_base_bdevs_discovered": 3, 00:09:29.633 "num_base_bdevs_operational": 3, 00:09:29.633 "base_bdevs_list": [ 00:09:29.633 { 00:09:29.633 "name": "BaseBdev1", 00:09:29.633 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:29.633 "is_configured": true, 00:09:29.633 "data_offset": 0, 00:09:29.633 "data_size": 65536 00:09:29.633 }, 00:09:29.633 { 00:09:29.633 "name": "BaseBdev2", 00:09:29.633 "uuid": "2514c8d5-bc66-4d24-ac63-4858f5033ba9", 00:09:29.633 "is_configured": true, 00:09:29.633 "data_offset": 0, 00:09:29.633 "data_size": 65536 00:09:29.633 }, 00:09:29.633 { 00:09:29.633 "name": "BaseBdev3", 00:09:29.633 "uuid": "b1cb77e0-4d4d-45e2-974d-a93e660ff1dc", 00:09:29.633 "is_configured": true, 00:09:29.633 "data_offset": 0, 00:09:29.633 "data_size": 65536 00:09:29.633 } 00:09:29.633 ] 00:09:29.633 }' 00:09:29.633 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.633 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:29.892 [2024-10-29 10:58:35.377412] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:29.892 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.151 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:30.151 "name": "Existed_Raid", 00:09:30.151 "aliases": [ 00:09:30.151 "0ceb04a2-b815-4dd8-9132-6f956ba17043" 00:09:30.151 ], 00:09:30.151 "product_name": "Raid Volume", 00:09:30.151 "block_size": 512, 00:09:30.151 "num_blocks": 65536, 00:09:30.151 "uuid": "0ceb04a2-b815-4dd8-9132-6f956ba17043", 00:09:30.151 "assigned_rate_limits": { 00:09:30.151 "rw_ios_per_sec": 0, 00:09:30.151 "rw_mbytes_per_sec": 0, 00:09:30.151 "r_mbytes_per_sec": 0, 00:09:30.151 "w_mbytes_per_sec": 0 00:09:30.151 }, 00:09:30.151 "claimed": false, 00:09:30.151 "zoned": false, 00:09:30.151 "supported_io_types": { 00:09:30.151 "read": true, 00:09:30.151 "write": true, 00:09:30.152 "unmap": false, 00:09:30.152 "flush": false, 00:09:30.152 "reset": true, 00:09:30.152 "nvme_admin": false, 00:09:30.152 "nvme_io": false, 00:09:30.152 "nvme_io_md": false, 00:09:30.152 "write_zeroes": true, 00:09:30.152 "zcopy": false, 00:09:30.152 "get_zone_info": false, 00:09:30.152 "zone_management": false, 00:09:30.152 "zone_append": false, 00:09:30.152 "compare": false, 00:09:30.152 "compare_and_write": false, 00:09:30.152 "abort": false, 00:09:30.152 "seek_hole": false, 00:09:30.152 "seek_data": false, 00:09:30.152 "copy": false, 00:09:30.152 "nvme_iov_md": false 00:09:30.152 }, 00:09:30.152 "memory_domains": [ 00:09:30.152 { 00:09:30.152 "dma_device_id": "system", 00:09:30.152 "dma_device_type": 1 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.152 "dma_device_type": 2 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "dma_device_id": "system", 00:09:30.152 "dma_device_type": 1 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.152 "dma_device_type": 2 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "dma_device_id": "system", 00:09:30.152 "dma_device_type": 1 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.152 "dma_device_type": 2 00:09:30.152 } 00:09:30.152 ], 00:09:30.152 "driver_specific": { 00:09:30.152 "raid": { 00:09:30.152 "uuid": "0ceb04a2-b815-4dd8-9132-6f956ba17043", 00:09:30.152 "strip_size_kb": 0, 00:09:30.152 "state": "online", 00:09:30.152 "raid_level": "raid1", 00:09:30.152 "superblock": false, 00:09:30.152 "num_base_bdevs": 3, 00:09:30.152 "num_base_bdevs_discovered": 3, 00:09:30.152 "num_base_bdevs_operational": 3, 00:09:30.152 "base_bdevs_list": [ 00:09:30.152 { 00:09:30.152 "name": "BaseBdev1", 00:09:30.152 "uuid": "15c556fa-79b3-4f7c-99a0-41f7100b22c7", 00:09:30.152 "is_configured": true, 00:09:30.152 "data_offset": 0, 00:09:30.152 "data_size": 65536 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "name": "BaseBdev2", 00:09:30.152 "uuid": "2514c8d5-bc66-4d24-ac63-4858f5033ba9", 00:09:30.152 "is_configured": true, 00:09:30.152 "data_offset": 0, 00:09:30.152 "data_size": 65536 00:09:30.152 }, 00:09:30.152 { 00:09:30.152 "name": "BaseBdev3", 00:09:30.152 "uuid": "b1cb77e0-4d4d-45e2-974d-a93e660ff1dc", 00:09:30.152 "is_configured": true, 00:09:30.152 "data_offset": 0, 00:09:30.152 "data_size": 65536 00:09:30.152 } 00:09:30.152 ] 00:09:30.152 } 00:09:30.152 } 00:09:30.152 }' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:30.152 BaseBdev2 00:09:30.152 BaseBdev3' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.152 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.152 [2024-10-29 10:58:35.648634] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.411 "name": "Existed_Raid", 00:09:30.411 "uuid": "0ceb04a2-b815-4dd8-9132-6f956ba17043", 00:09:30.411 "strip_size_kb": 0, 00:09:30.411 "state": "online", 00:09:30.411 "raid_level": "raid1", 00:09:30.411 "superblock": false, 00:09:30.411 "num_base_bdevs": 3, 00:09:30.411 "num_base_bdevs_discovered": 2, 00:09:30.411 "num_base_bdevs_operational": 2, 00:09:30.411 "base_bdevs_list": [ 00:09:30.411 { 00:09:30.411 "name": null, 00:09:30.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.411 "is_configured": false, 00:09:30.411 "data_offset": 0, 00:09:30.411 "data_size": 65536 00:09:30.411 }, 00:09:30.411 { 00:09:30.411 "name": "BaseBdev2", 00:09:30.411 "uuid": "2514c8d5-bc66-4d24-ac63-4858f5033ba9", 00:09:30.411 "is_configured": true, 00:09:30.411 "data_offset": 0, 00:09:30.411 "data_size": 65536 00:09:30.411 }, 00:09:30.411 { 00:09:30.411 "name": "BaseBdev3", 00:09:30.411 "uuid": "b1cb77e0-4d4d-45e2-974d-a93e660ff1dc", 00:09:30.411 "is_configured": true, 00:09:30.411 "data_offset": 0, 00:09:30.411 "data_size": 65536 00:09:30.411 } 00:09:30.411 ] 00:09:30.411 }' 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.411 10:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.670 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.670 [2024-10-29 10:58:36.156812] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:30.930 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.930 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:30.930 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.930 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.930 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 [2024-10-29 10:58:36.237194] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:30.931 [2024-10-29 10:58:36.237370] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:30.931 [2024-10-29 10:58:36.257529] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:30.931 [2024-10-29 10:58:36.257631] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:30.931 [2024-10-29 10:58:36.257678] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 BaseBdev2 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 [ 00:09:30.931 { 00:09:30.931 "name": "BaseBdev2", 00:09:30.931 "aliases": [ 00:09:30.931 "cdb1ee28-7f82-4851-af62-b0890a45def8" 00:09:30.931 ], 00:09:30.931 "product_name": "Malloc disk", 00:09:30.931 "block_size": 512, 00:09:30.931 "num_blocks": 65536, 00:09:30.931 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:30.931 "assigned_rate_limits": { 00:09:30.931 "rw_ios_per_sec": 0, 00:09:30.931 "rw_mbytes_per_sec": 0, 00:09:30.931 "r_mbytes_per_sec": 0, 00:09:30.931 "w_mbytes_per_sec": 0 00:09:30.931 }, 00:09:30.931 "claimed": false, 00:09:30.931 "zoned": false, 00:09:30.931 "supported_io_types": { 00:09:30.931 "read": true, 00:09:30.931 "write": true, 00:09:30.931 "unmap": true, 00:09:30.931 "flush": true, 00:09:30.931 "reset": true, 00:09:30.931 "nvme_admin": false, 00:09:30.931 "nvme_io": false, 00:09:30.931 "nvme_io_md": false, 00:09:30.931 "write_zeroes": true, 00:09:30.931 "zcopy": true, 00:09:30.931 "get_zone_info": false, 00:09:30.931 "zone_management": false, 00:09:30.931 "zone_append": false, 00:09:30.931 "compare": false, 00:09:30.931 "compare_and_write": false, 00:09:30.931 "abort": true, 00:09:30.931 "seek_hole": false, 00:09:30.931 "seek_data": false, 00:09:30.931 "copy": true, 00:09:30.931 "nvme_iov_md": false 00:09:30.931 }, 00:09:30.931 "memory_domains": [ 00:09:30.931 { 00:09:30.931 "dma_device_id": "system", 00:09:30.931 "dma_device_type": 1 00:09:30.931 }, 00:09:30.931 { 00:09:30.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.931 "dma_device_type": 2 00:09:30.931 } 00:09:30.931 ], 00:09:30.931 "driver_specific": {} 00:09:30.931 } 00:09:30.931 ] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 BaseBdev3 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.931 [ 00:09:30.931 { 00:09:30.931 "name": "BaseBdev3", 00:09:30.931 "aliases": [ 00:09:30.931 "18ba0f17-d324-481d-8d9f-37da51be8d6a" 00:09:30.931 ], 00:09:30.931 "product_name": "Malloc disk", 00:09:30.931 "block_size": 512, 00:09:30.931 "num_blocks": 65536, 00:09:30.931 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:30.931 "assigned_rate_limits": { 00:09:30.931 "rw_ios_per_sec": 0, 00:09:30.931 "rw_mbytes_per_sec": 0, 00:09:30.931 "r_mbytes_per_sec": 0, 00:09:30.931 "w_mbytes_per_sec": 0 00:09:30.931 }, 00:09:30.931 "claimed": false, 00:09:30.931 "zoned": false, 00:09:30.931 "supported_io_types": { 00:09:30.931 "read": true, 00:09:30.931 "write": true, 00:09:30.931 "unmap": true, 00:09:30.931 "flush": true, 00:09:30.931 "reset": true, 00:09:30.931 "nvme_admin": false, 00:09:30.931 "nvme_io": false, 00:09:30.931 "nvme_io_md": false, 00:09:30.931 "write_zeroes": true, 00:09:30.931 "zcopy": true, 00:09:30.931 "get_zone_info": false, 00:09:30.931 "zone_management": false, 00:09:30.931 "zone_append": false, 00:09:30.931 "compare": false, 00:09:30.931 "compare_and_write": false, 00:09:30.931 "abort": true, 00:09:30.931 "seek_hole": false, 00:09:30.931 "seek_data": false, 00:09:30.931 "copy": true, 00:09:30.931 "nvme_iov_md": false 00:09:30.931 }, 00:09:30.931 "memory_domains": [ 00:09:30.931 { 00:09:30.931 "dma_device_id": "system", 00:09:30.931 "dma_device_type": 1 00:09:30.931 }, 00:09:30.931 { 00:09:30.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.931 "dma_device_type": 2 00:09:30.931 } 00:09:30.931 ], 00:09:30.931 "driver_specific": {} 00:09:30.931 } 00:09:30.931 ] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.931 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.191 [2024-10-29 10:58:36.431215] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:31.191 [2024-10-29 10:58:36.431337] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:31.191 [2024-10-29 10:58:36.431392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:31.191 [2024-10-29 10:58:36.433617] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.191 "name": "Existed_Raid", 00:09:31.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.191 "strip_size_kb": 0, 00:09:31.191 "state": "configuring", 00:09:31.191 "raid_level": "raid1", 00:09:31.191 "superblock": false, 00:09:31.191 "num_base_bdevs": 3, 00:09:31.191 "num_base_bdevs_discovered": 2, 00:09:31.191 "num_base_bdevs_operational": 3, 00:09:31.191 "base_bdevs_list": [ 00:09:31.191 { 00:09:31.191 "name": "BaseBdev1", 00:09:31.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.191 "is_configured": false, 00:09:31.191 "data_offset": 0, 00:09:31.191 "data_size": 0 00:09:31.191 }, 00:09:31.191 { 00:09:31.191 "name": "BaseBdev2", 00:09:31.191 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:31.191 "is_configured": true, 00:09:31.191 "data_offset": 0, 00:09:31.191 "data_size": 65536 00:09:31.191 }, 00:09:31.191 { 00:09:31.191 "name": "BaseBdev3", 00:09:31.191 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:31.191 "is_configured": true, 00:09:31.191 "data_offset": 0, 00:09:31.191 "data_size": 65536 00:09:31.191 } 00:09:31.191 ] 00:09:31.191 }' 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.191 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.450 [2024-10-29 10:58:36.930448] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.450 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.709 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.709 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.709 "name": "Existed_Raid", 00:09:31.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.709 "strip_size_kb": 0, 00:09:31.709 "state": "configuring", 00:09:31.709 "raid_level": "raid1", 00:09:31.709 "superblock": false, 00:09:31.709 "num_base_bdevs": 3, 00:09:31.709 "num_base_bdevs_discovered": 1, 00:09:31.709 "num_base_bdevs_operational": 3, 00:09:31.709 "base_bdevs_list": [ 00:09:31.709 { 00:09:31.709 "name": "BaseBdev1", 00:09:31.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.709 "is_configured": false, 00:09:31.709 "data_offset": 0, 00:09:31.709 "data_size": 0 00:09:31.709 }, 00:09:31.709 { 00:09:31.709 "name": null, 00:09:31.709 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:31.709 "is_configured": false, 00:09:31.709 "data_offset": 0, 00:09:31.709 "data_size": 65536 00:09:31.709 }, 00:09:31.709 { 00:09:31.709 "name": "BaseBdev3", 00:09:31.709 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:31.709 "is_configured": true, 00:09:31.709 "data_offset": 0, 00:09:31.709 "data_size": 65536 00:09:31.709 } 00:09:31.709 ] 00:09:31.709 }' 00:09:31.709 10:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.709 10:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.967 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.968 [2024-10-29 10:58:37.386295] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:31.968 BaseBdev1 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.968 [ 00:09:31.968 { 00:09:31.968 "name": "BaseBdev1", 00:09:31.968 "aliases": [ 00:09:31.968 "43083025-ec28-4fe8-83f5-ec58c98cee6e" 00:09:31.968 ], 00:09:31.968 "product_name": "Malloc disk", 00:09:31.968 "block_size": 512, 00:09:31.968 "num_blocks": 65536, 00:09:31.968 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:31.968 "assigned_rate_limits": { 00:09:31.968 "rw_ios_per_sec": 0, 00:09:31.968 "rw_mbytes_per_sec": 0, 00:09:31.968 "r_mbytes_per_sec": 0, 00:09:31.968 "w_mbytes_per_sec": 0 00:09:31.968 }, 00:09:31.968 "claimed": true, 00:09:31.968 "claim_type": "exclusive_write", 00:09:31.968 "zoned": false, 00:09:31.968 "supported_io_types": { 00:09:31.968 "read": true, 00:09:31.968 "write": true, 00:09:31.968 "unmap": true, 00:09:31.968 "flush": true, 00:09:31.968 "reset": true, 00:09:31.968 "nvme_admin": false, 00:09:31.968 "nvme_io": false, 00:09:31.968 "nvme_io_md": false, 00:09:31.968 "write_zeroes": true, 00:09:31.968 "zcopy": true, 00:09:31.968 "get_zone_info": false, 00:09:31.968 "zone_management": false, 00:09:31.968 "zone_append": false, 00:09:31.968 "compare": false, 00:09:31.968 "compare_and_write": false, 00:09:31.968 "abort": true, 00:09:31.968 "seek_hole": false, 00:09:31.968 "seek_data": false, 00:09:31.968 "copy": true, 00:09:31.968 "nvme_iov_md": false 00:09:31.968 }, 00:09:31.968 "memory_domains": [ 00:09:31.968 { 00:09:31.968 "dma_device_id": "system", 00:09:31.968 "dma_device_type": 1 00:09:31.968 }, 00:09:31.968 { 00:09:31.968 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:31.968 "dma_device_type": 2 00:09:31.968 } 00:09:31.968 ], 00:09:31.968 "driver_specific": {} 00:09:31.968 } 00:09:31.968 ] 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.968 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.226 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.226 "name": "Existed_Raid", 00:09:32.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.226 "strip_size_kb": 0, 00:09:32.226 "state": "configuring", 00:09:32.226 "raid_level": "raid1", 00:09:32.226 "superblock": false, 00:09:32.226 "num_base_bdevs": 3, 00:09:32.226 "num_base_bdevs_discovered": 2, 00:09:32.226 "num_base_bdevs_operational": 3, 00:09:32.226 "base_bdevs_list": [ 00:09:32.226 { 00:09:32.226 "name": "BaseBdev1", 00:09:32.226 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:32.226 "is_configured": true, 00:09:32.226 "data_offset": 0, 00:09:32.226 "data_size": 65536 00:09:32.226 }, 00:09:32.226 { 00:09:32.226 "name": null, 00:09:32.226 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:32.226 "is_configured": false, 00:09:32.226 "data_offset": 0, 00:09:32.226 "data_size": 65536 00:09:32.226 }, 00:09:32.226 { 00:09:32.226 "name": "BaseBdev3", 00:09:32.226 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:32.226 "is_configured": true, 00:09:32.226 "data_offset": 0, 00:09:32.226 "data_size": 65536 00:09:32.226 } 00:09:32.226 ] 00:09:32.226 }' 00:09:32.226 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.226 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.483 [2024-10-29 10:58:37.893530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.483 "name": "Existed_Raid", 00:09:32.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.483 "strip_size_kb": 0, 00:09:32.483 "state": "configuring", 00:09:32.483 "raid_level": "raid1", 00:09:32.483 "superblock": false, 00:09:32.483 "num_base_bdevs": 3, 00:09:32.483 "num_base_bdevs_discovered": 1, 00:09:32.483 "num_base_bdevs_operational": 3, 00:09:32.483 "base_bdevs_list": [ 00:09:32.483 { 00:09:32.483 "name": "BaseBdev1", 00:09:32.483 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:32.483 "is_configured": true, 00:09:32.483 "data_offset": 0, 00:09:32.483 "data_size": 65536 00:09:32.483 }, 00:09:32.483 { 00:09:32.483 "name": null, 00:09:32.483 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:32.483 "is_configured": false, 00:09:32.483 "data_offset": 0, 00:09:32.483 "data_size": 65536 00:09:32.483 }, 00:09:32.483 { 00:09:32.483 "name": null, 00:09:32.483 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:32.483 "is_configured": false, 00:09:32.483 "data_offset": 0, 00:09:32.483 "data_size": 65536 00:09:32.483 } 00:09:32.483 ] 00:09:32.483 }' 00:09:32.483 10:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.484 10:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.050 [2024-10-29 10:58:38.380715] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.050 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.050 "name": "Existed_Raid", 00:09:33.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.050 "strip_size_kb": 0, 00:09:33.050 "state": "configuring", 00:09:33.050 "raid_level": "raid1", 00:09:33.050 "superblock": false, 00:09:33.050 "num_base_bdevs": 3, 00:09:33.050 "num_base_bdevs_discovered": 2, 00:09:33.050 "num_base_bdevs_operational": 3, 00:09:33.051 "base_bdevs_list": [ 00:09:33.051 { 00:09:33.051 "name": "BaseBdev1", 00:09:33.051 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:33.051 "is_configured": true, 00:09:33.051 "data_offset": 0, 00:09:33.051 "data_size": 65536 00:09:33.051 }, 00:09:33.051 { 00:09:33.051 "name": null, 00:09:33.051 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:33.051 "is_configured": false, 00:09:33.051 "data_offset": 0, 00:09:33.051 "data_size": 65536 00:09:33.051 }, 00:09:33.051 { 00:09:33.051 "name": "BaseBdev3", 00:09:33.051 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:33.051 "is_configured": true, 00:09:33.051 "data_offset": 0, 00:09:33.051 "data_size": 65536 00:09:33.051 } 00:09:33.051 ] 00:09:33.051 }' 00:09:33.051 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.051 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.309 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.309 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.309 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.309 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:33.309 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.570 [2024-10-29 10:58:38.848054] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.570 "name": "Existed_Raid", 00:09:33.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.570 "strip_size_kb": 0, 00:09:33.570 "state": "configuring", 00:09:33.570 "raid_level": "raid1", 00:09:33.570 "superblock": false, 00:09:33.570 "num_base_bdevs": 3, 00:09:33.570 "num_base_bdevs_discovered": 1, 00:09:33.570 "num_base_bdevs_operational": 3, 00:09:33.570 "base_bdevs_list": [ 00:09:33.570 { 00:09:33.570 "name": null, 00:09:33.570 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:33.570 "is_configured": false, 00:09:33.570 "data_offset": 0, 00:09:33.570 "data_size": 65536 00:09:33.570 }, 00:09:33.570 { 00:09:33.570 "name": null, 00:09:33.570 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:33.570 "is_configured": false, 00:09:33.570 "data_offset": 0, 00:09:33.570 "data_size": 65536 00:09:33.570 }, 00:09:33.570 { 00:09:33.570 "name": "BaseBdev3", 00:09:33.570 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:33.570 "is_configured": true, 00:09:33.570 "data_offset": 0, 00:09:33.570 "data_size": 65536 00:09:33.570 } 00:09:33.570 ] 00:09:33.570 }' 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.570 10:58:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.830 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:33.830 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.830 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.830 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.830 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.089 [2024-10-29 10:58:39.335645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.089 "name": "Existed_Raid", 00:09:34.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.089 "strip_size_kb": 0, 00:09:34.089 "state": "configuring", 00:09:34.089 "raid_level": "raid1", 00:09:34.089 "superblock": false, 00:09:34.089 "num_base_bdevs": 3, 00:09:34.089 "num_base_bdevs_discovered": 2, 00:09:34.089 "num_base_bdevs_operational": 3, 00:09:34.089 "base_bdevs_list": [ 00:09:34.089 { 00:09:34.089 "name": null, 00:09:34.089 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:34.089 "is_configured": false, 00:09:34.089 "data_offset": 0, 00:09:34.089 "data_size": 65536 00:09:34.089 }, 00:09:34.089 { 00:09:34.089 "name": "BaseBdev2", 00:09:34.089 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:34.089 "is_configured": true, 00:09:34.089 "data_offset": 0, 00:09:34.089 "data_size": 65536 00:09:34.089 }, 00:09:34.089 { 00:09:34.089 "name": "BaseBdev3", 00:09:34.089 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:34.089 "is_configured": true, 00:09:34.089 "data_offset": 0, 00:09:34.089 "data_size": 65536 00:09:34.089 } 00:09:34.089 ] 00:09:34.089 }' 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.089 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 43083025-ec28-4fe8-83f5-ec58c98cee6e 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.349 [2024-10-29 10:58:39.835782] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:34.349 [2024-10-29 10:58:39.835918] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:34.349 [2024-10-29 10:58:39.835944] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:34.349 [2024-10-29 10:58:39.836298] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:34.349 [2024-10-29 10:58:39.836507] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:34.349 [2024-10-29 10:58:39.836558] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:34.349 [2024-10-29 10:58:39.836866] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.349 NewBaseBdev 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.349 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.608 [ 00:09:34.608 { 00:09:34.608 "name": "NewBaseBdev", 00:09:34.608 "aliases": [ 00:09:34.608 "43083025-ec28-4fe8-83f5-ec58c98cee6e" 00:09:34.608 ], 00:09:34.608 "product_name": "Malloc disk", 00:09:34.608 "block_size": 512, 00:09:34.608 "num_blocks": 65536, 00:09:34.608 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:34.608 "assigned_rate_limits": { 00:09:34.608 "rw_ios_per_sec": 0, 00:09:34.608 "rw_mbytes_per_sec": 0, 00:09:34.608 "r_mbytes_per_sec": 0, 00:09:34.608 "w_mbytes_per_sec": 0 00:09:34.608 }, 00:09:34.608 "claimed": true, 00:09:34.608 "claim_type": "exclusive_write", 00:09:34.608 "zoned": false, 00:09:34.608 "supported_io_types": { 00:09:34.608 "read": true, 00:09:34.608 "write": true, 00:09:34.608 "unmap": true, 00:09:34.608 "flush": true, 00:09:34.608 "reset": true, 00:09:34.608 "nvme_admin": false, 00:09:34.608 "nvme_io": false, 00:09:34.608 "nvme_io_md": false, 00:09:34.608 "write_zeroes": true, 00:09:34.608 "zcopy": true, 00:09:34.608 "get_zone_info": false, 00:09:34.608 "zone_management": false, 00:09:34.608 "zone_append": false, 00:09:34.608 "compare": false, 00:09:34.608 "compare_and_write": false, 00:09:34.608 "abort": true, 00:09:34.608 "seek_hole": false, 00:09:34.608 "seek_data": false, 00:09:34.608 "copy": true, 00:09:34.608 "nvme_iov_md": false 00:09:34.608 }, 00:09:34.608 "memory_domains": [ 00:09:34.608 { 00:09:34.608 "dma_device_id": "system", 00:09:34.608 "dma_device_type": 1 00:09:34.608 }, 00:09:34.608 { 00:09:34.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.608 "dma_device_type": 2 00:09:34.608 } 00:09:34.608 ], 00:09:34.608 "driver_specific": {} 00:09:34.608 } 00:09:34.608 ] 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.608 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.609 "name": "Existed_Raid", 00:09:34.609 "uuid": "944a21f4-1c22-44c2-a0c9-7ad5adebcba7", 00:09:34.609 "strip_size_kb": 0, 00:09:34.609 "state": "online", 00:09:34.609 "raid_level": "raid1", 00:09:34.609 "superblock": false, 00:09:34.609 "num_base_bdevs": 3, 00:09:34.609 "num_base_bdevs_discovered": 3, 00:09:34.609 "num_base_bdevs_operational": 3, 00:09:34.609 "base_bdevs_list": [ 00:09:34.609 { 00:09:34.609 "name": "NewBaseBdev", 00:09:34.609 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:34.609 "is_configured": true, 00:09:34.609 "data_offset": 0, 00:09:34.609 "data_size": 65536 00:09:34.609 }, 00:09:34.609 { 00:09:34.609 "name": "BaseBdev2", 00:09:34.609 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:34.609 "is_configured": true, 00:09:34.609 "data_offset": 0, 00:09:34.609 "data_size": 65536 00:09:34.609 }, 00:09:34.609 { 00:09:34.609 "name": "BaseBdev3", 00:09:34.609 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:34.609 "is_configured": true, 00:09:34.609 "data_offset": 0, 00:09:34.609 "data_size": 65536 00:09:34.609 } 00:09:34.609 ] 00:09:34.609 }' 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.609 10:58:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:34.868 [2024-10-29 10:58:40.295513] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.868 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:34.868 "name": "Existed_Raid", 00:09:34.868 "aliases": [ 00:09:34.868 "944a21f4-1c22-44c2-a0c9-7ad5adebcba7" 00:09:34.868 ], 00:09:34.868 "product_name": "Raid Volume", 00:09:34.868 "block_size": 512, 00:09:34.868 "num_blocks": 65536, 00:09:34.868 "uuid": "944a21f4-1c22-44c2-a0c9-7ad5adebcba7", 00:09:34.868 "assigned_rate_limits": { 00:09:34.868 "rw_ios_per_sec": 0, 00:09:34.868 "rw_mbytes_per_sec": 0, 00:09:34.868 "r_mbytes_per_sec": 0, 00:09:34.868 "w_mbytes_per_sec": 0 00:09:34.868 }, 00:09:34.868 "claimed": false, 00:09:34.868 "zoned": false, 00:09:34.868 "supported_io_types": { 00:09:34.868 "read": true, 00:09:34.868 "write": true, 00:09:34.868 "unmap": false, 00:09:34.868 "flush": false, 00:09:34.868 "reset": true, 00:09:34.868 "nvme_admin": false, 00:09:34.868 "nvme_io": false, 00:09:34.868 "nvme_io_md": false, 00:09:34.868 "write_zeroes": true, 00:09:34.868 "zcopy": false, 00:09:34.868 "get_zone_info": false, 00:09:34.868 "zone_management": false, 00:09:34.868 "zone_append": false, 00:09:34.868 "compare": false, 00:09:34.868 "compare_and_write": false, 00:09:34.868 "abort": false, 00:09:34.869 "seek_hole": false, 00:09:34.869 "seek_data": false, 00:09:34.869 "copy": false, 00:09:34.869 "nvme_iov_md": false 00:09:34.869 }, 00:09:34.869 "memory_domains": [ 00:09:34.869 { 00:09:34.869 "dma_device_id": "system", 00:09:34.869 "dma_device_type": 1 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.869 "dma_device_type": 2 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "dma_device_id": "system", 00:09:34.869 "dma_device_type": 1 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.869 "dma_device_type": 2 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "dma_device_id": "system", 00:09:34.869 "dma_device_type": 1 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.869 "dma_device_type": 2 00:09:34.869 } 00:09:34.869 ], 00:09:34.869 "driver_specific": { 00:09:34.869 "raid": { 00:09:34.869 "uuid": "944a21f4-1c22-44c2-a0c9-7ad5adebcba7", 00:09:34.869 "strip_size_kb": 0, 00:09:34.869 "state": "online", 00:09:34.869 "raid_level": "raid1", 00:09:34.869 "superblock": false, 00:09:34.869 "num_base_bdevs": 3, 00:09:34.869 "num_base_bdevs_discovered": 3, 00:09:34.869 "num_base_bdevs_operational": 3, 00:09:34.869 "base_bdevs_list": [ 00:09:34.869 { 00:09:34.869 "name": "NewBaseBdev", 00:09:34.869 "uuid": "43083025-ec28-4fe8-83f5-ec58c98cee6e", 00:09:34.869 "is_configured": true, 00:09:34.869 "data_offset": 0, 00:09:34.869 "data_size": 65536 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "name": "BaseBdev2", 00:09:34.869 "uuid": "cdb1ee28-7f82-4851-af62-b0890a45def8", 00:09:34.869 "is_configured": true, 00:09:34.869 "data_offset": 0, 00:09:34.869 "data_size": 65536 00:09:34.869 }, 00:09:34.869 { 00:09:34.869 "name": "BaseBdev3", 00:09:34.869 "uuid": "18ba0f17-d324-481d-8d9f-37da51be8d6a", 00:09:34.869 "is_configured": true, 00:09:34.869 "data_offset": 0, 00:09:34.869 "data_size": 65536 00:09:34.869 } 00:09:34.869 ] 00:09:34.869 } 00:09:34.869 } 00:09:34.869 }' 00:09:34.869 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:34.869 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:34.869 BaseBdev2 00:09:34.869 BaseBdev3' 00:09:34.869 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.129 [2024-10-29 10:58:40.546632] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:35.129 [2024-10-29 10:58:40.546680] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:35.129 [2024-10-29 10:58:40.546789] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:35.129 [2024-10-29 10:58:40.547077] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:35.129 [2024-10-29 10:58:40.547090] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78594 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 78594 ']' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 78594 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 78594 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 78594' 00:09:35.129 killing process with pid 78594 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 78594 00:09:35.129 [2024-10-29 10:58:40.596797] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:35.129 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 78594 00:09:35.389 [2024-10-29 10:58:40.655940] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:35.649 10:58:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:35.649 00:09:35.649 real 0m8.869s 00:09:35.649 user 0m14.879s 00:09:35.649 sys 0m1.782s 00:09:35.649 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:35.649 ************************************ 00:09:35.649 END TEST raid_state_function_test 00:09:35.649 ************************************ 00:09:35.649 10:58:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.650 10:58:41 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:09:35.650 10:58:41 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:35.650 10:58:41 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:35.650 10:58:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:35.650 ************************************ 00:09:35.650 START TEST raid_state_function_test_sb 00:09:35.650 ************************************ 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 3 true 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=79203 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 79203' 00:09:35.650 Process raid pid: 79203 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 79203 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 79203 ']' 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:35.650 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:35.650 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.909 [2024-10-29 10:58:41.149081] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:35.909 [2024-10-29 10:58:41.149299] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:35.909 [2024-10-29 10:58:41.321089] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:35.909 [2024-10-29 10:58:41.363818] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:36.168 [2024-10-29 10:58:41.441040] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:36.168 [2024-10-29 10:58:41.441205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.737 [2024-10-29 10:58:41.972856] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:36.737 [2024-10-29 10:58:41.972979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:36.737 [2024-10-29 10:58:41.973029] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:36.737 [2024-10-29 10:58:41.973107] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:36.737 [2024-10-29 10:58:41.973134] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:36.737 [2024-10-29 10:58:41.973178] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.737 10:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.737 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.737 "name": "Existed_Raid", 00:09:36.737 "uuid": "628e380d-8caf-4e85-b048-a088c96c7d1f", 00:09:36.737 "strip_size_kb": 0, 00:09:36.737 "state": "configuring", 00:09:36.737 "raid_level": "raid1", 00:09:36.737 "superblock": true, 00:09:36.737 "num_base_bdevs": 3, 00:09:36.737 "num_base_bdevs_discovered": 0, 00:09:36.737 "num_base_bdevs_operational": 3, 00:09:36.737 "base_bdevs_list": [ 00:09:36.737 { 00:09:36.737 "name": "BaseBdev1", 00:09:36.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.737 "is_configured": false, 00:09:36.737 "data_offset": 0, 00:09:36.737 "data_size": 0 00:09:36.737 }, 00:09:36.737 { 00:09:36.737 "name": "BaseBdev2", 00:09:36.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.737 "is_configured": false, 00:09:36.737 "data_offset": 0, 00:09:36.737 "data_size": 0 00:09:36.737 }, 00:09:36.737 { 00:09:36.737 "name": "BaseBdev3", 00:09:36.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.737 "is_configured": false, 00:09:36.737 "data_offset": 0, 00:09:36.737 "data_size": 0 00:09:36.737 } 00:09:36.737 ] 00:09:36.737 }' 00:09:36.737 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.737 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 [2024-10-29 10:58:42.360164] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:36.997 [2024-10-29 10:58:42.360230] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 [2024-10-29 10:58:42.368143] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:36.997 [2024-10-29 10:58:42.368208] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:36.997 [2024-10-29 10:58:42.368218] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:36.997 [2024-10-29 10:58:42.368229] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:36.997 [2024-10-29 10:58:42.368236] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:36.997 [2024-10-29 10:58:42.368246] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 [2024-10-29 10:58:42.391557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:36.997 BaseBdev1 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.997 [ 00:09:36.997 { 00:09:36.997 "name": "BaseBdev1", 00:09:36.997 "aliases": [ 00:09:36.997 "937939e1-96bf-477a-a04f-08937cdd8faf" 00:09:36.997 ], 00:09:36.997 "product_name": "Malloc disk", 00:09:36.997 "block_size": 512, 00:09:36.997 "num_blocks": 65536, 00:09:36.997 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:36.997 "assigned_rate_limits": { 00:09:36.997 "rw_ios_per_sec": 0, 00:09:36.997 "rw_mbytes_per_sec": 0, 00:09:36.997 "r_mbytes_per_sec": 0, 00:09:36.997 "w_mbytes_per_sec": 0 00:09:36.997 }, 00:09:36.997 "claimed": true, 00:09:36.997 "claim_type": "exclusive_write", 00:09:36.997 "zoned": false, 00:09:36.997 "supported_io_types": { 00:09:36.997 "read": true, 00:09:36.997 "write": true, 00:09:36.997 "unmap": true, 00:09:36.997 "flush": true, 00:09:36.997 "reset": true, 00:09:36.997 "nvme_admin": false, 00:09:36.997 "nvme_io": false, 00:09:36.997 "nvme_io_md": false, 00:09:36.997 "write_zeroes": true, 00:09:36.997 "zcopy": true, 00:09:36.997 "get_zone_info": false, 00:09:36.997 "zone_management": false, 00:09:36.997 "zone_append": false, 00:09:36.997 "compare": false, 00:09:36.997 "compare_and_write": false, 00:09:36.997 "abort": true, 00:09:36.997 "seek_hole": false, 00:09:36.997 "seek_data": false, 00:09:36.997 "copy": true, 00:09:36.997 "nvme_iov_md": false 00:09:36.997 }, 00:09:36.997 "memory_domains": [ 00:09:36.997 { 00:09:36.997 "dma_device_id": "system", 00:09:36.997 "dma_device_type": 1 00:09:36.997 }, 00:09:36.997 { 00:09:36.997 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.997 "dma_device_type": 2 00:09:36.997 } 00:09:36.997 ], 00:09:36.997 "driver_specific": {} 00:09:36.997 } 00:09:36.997 ] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.997 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.998 "name": "Existed_Raid", 00:09:36.998 "uuid": "78d40d54-7980-49e7-9f69-b69eefb99030", 00:09:36.998 "strip_size_kb": 0, 00:09:36.998 "state": "configuring", 00:09:36.998 "raid_level": "raid1", 00:09:36.998 "superblock": true, 00:09:36.998 "num_base_bdevs": 3, 00:09:36.998 "num_base_bdevs_discovered": 1, 00:09:36.998 "num_base_bdevs_operational": 3, 00:09:36.998 "base_bdevs_list": [ 00:09:36.998 { 00:09:36.998 "name": "BaseBdev1", 00:09:36.998 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:36.998 "is_configured": true, 00:09:36.998 "data_offset": 2048, 00:09:36.998 "data_size": 63488 00:09:36.998 }, 00:09:36.998 { 00:09:36.998 "name": "BaseBdev2", 00:09:36.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.998 "is_configured": false, 00:09:36.998 "data_offset": 0, 00:09:36.998 "data_size": 0 00:09:36.998 }, 00:09:36.998 { 00:09:36.998 "name": "BaseBdev3", 00:09:36.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.998 "is_configured": false, 00:09:36.998 "data_offset": 0, 00:09:36.998 "data_size": 0 00:09:36.998 } 00:09:36.998 ] 00:09:36.998 }' 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.998 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.567 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:37.567 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.567 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.567 [2024-10-29 10:58:42.866833] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:37.567 [2024-10-29 10:58:42.867026] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:37.567 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.568 [2024-10-29 10:58:42.878882] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:37.568 [2024-10-29 10:58:42.881217] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:37.568 [2024-10-29 10:58:42.881271] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:37.568 [2024-10-29 10:58:42.881282] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:37.568 [2024-10-29 10:58:42.881291] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.568 "name": "Existed_Raid", 00:09:37.568 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:37.568 "strip_size_kb": 0, 00:09:37.568 "state": "configuring", 00:09:37.568 "raid_level": "raid1", 00:09:37.568 "superblock": true, 00:09:37.568 "num_base_bdevs": 3, 00:09:37.568 "num_base_bdevs_discovered": 1, 00:09:37.568 "num_base_bdevs_operational": 3, 00:09:37.568 "base_bdevs_list": [ 00:09:37.568 { 00:09:37.568 "name": "BaseBdev1", 00:09:37.568 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:37.568 "is_configured": true, 00:09:37.568 "data_offset": 2048, 00:09:37.568 "data_size": 63488 00:09:37.568 }, 00:09:37.568 { 00:09:37.568 "name": "BaseBdev2", 00:09:37.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.568 "is_configured": false, 00:09:37.568 "data_offset": 0, 00:09:37.568 "data_size": 0 00:09:37.568 }, 00:09:37.568 { 00:09:37.568 "name": "BaseBdev3", 00:09:37.568 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.568 "is_configured": false, 00:09:37.568 "data_offset": 0, 00:09:37.568 "data_size": 0 00:09:37.568 } 00:09:37.568 ] 00:09:37.568 }' 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.568 10:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.828 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:37.828 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.828 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.828 [2024-10-29 10:58:43.326975] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:37.828 BaseBdev2 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.087 [ 00:09:38.087 { 00:09:38.087 "name": "BaseBdev2", 00:09:38.087 "aliases": [ 00:09:38.087 "06b18932-65ed-4c40-bb45-d2f8635cdaad" 00:09:38.087 ], 00:09:38.087 "product_name": "Malloc disk", 00:09:38.087 "block_size": 512, 00:09:38.087 "num_blocks": 65536, 00:09:38.087 "uuid": "06b18932-65ed-4c40-bb45-d2f8635cdaad", 00:09:38.087 "assigned_rate_limits": { 00:09:38.087 "rw_ios_per_sec": 0, 00:09:38.087 "rw_mbytes_per_sec": 0, 00:09:38.087 "r_mbytes_per_sec": 0, 00:09:38.087 "w_mbytes_per_sec": 0 00:09:38.087 }, 00:09:38.087 "claimed": true, 00:09:38.087 "claim_type": "exclusive_write", 00:09:38.087 "zoned": false, 00:09:38.087 "supported_io_types": { 00:09:38.087 "read": true, 00:09:38.087 "write": true, 00:09:38.087 "unmap": true, 00:09:38.087 "flush": true, 00:09:38.087 "reset": true, 00:09:38.087 "nvme_admin": false, 00:09:38.087 "nvme_io": false, 00:09:38.087 "nvme_io_md": false, 00:09:38.087 "write_zeroes": true, 00:09:38.087 "zcopy": true, 00:09:38.087 "get_zone_info": false, 00:09:38.087 "zone_management": false, 00:09:38.087 "zone_append": false, 00:09:38.087 "compare": false, 00:09:38.087 "compare_and_write": false, 00:09:38.087 "abort": true, 00:09:38.087 "seek_hole": false, 00:09:38.087 "seek_data": false, 00:09:38.087 "copy": true, 00:09:38.087 "nvme_iov_md": false 00:09:38.087 }, 00:09:38.087 "memory_domains": [ 00:09:38.087 { 00:09:38.087 "dma_device_id": "system", 00:09:38.087 "dma_device_type": 1 00:09:38.087 }, 00:09:38.087 { 00:09:38.087 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.087 "dma_device_type": 2 00:09:38.087 } 00:09:38.087 ], 00:09:38.087 "driver_specific": {} 00:09:38.087 } 00:09:38.087 ] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.087 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.087 "name": "Existed_Raid", 00:09:38.087 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:38.087 "strip_size_kb": 0, 00:09:38.087 "state": "configuring", 00:09:38.087 "raid_level": "raid1", 00:09:38.087 "superblock": true, 00:09:38.087 "num_base_bdevs": 3, 00:09:38.087 "num_base_bdevs_discovered": 2, 00:09:38.087 "num_base_bdevs_operational": 3, 00:09:38.087 "base_bdevs_list": [ 00:09:38.087 { 00:09:38.087 "name": "BaseBdev1", 00:09:38.087 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:38.087 "is_configured": true, 00:09:38.087 "data_offset": 2048, 00:09:38.087 "data_size": 63488 00:09:38.087 }, 00:09:38.087 { 00:09:38.087 "name": "BaseBdev2", 00:09:38.087 "uuid": "06b18932-65ed-4c40-bb45-d2f8635cdaad", 00:09:38.087 "is_configured": true, 00:09:38.087 "data_offset": 2048, 00:09:38.087 "data_size": 63488 00:09:38.087 }, 00:09:38.087 { 00:09:38.087 "name": "BaseBdev3", 00:09:38.088 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.088 "is_configured": false, 00:09:38.088 "data_offset": 0, 00:09:38.088 "data_size": 0 00:09:38.088 } 00:09:38.088 ] 00:09:38.088 }' 00:09:38.088 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.088 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.347 [2024-10-29 10:58:43.829550] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:38.347 [2024-10-29 10:58:43.829794] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:38.347 [2024-10-29 10:58:43.829822] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:38.347 BaseBdev3 00:09:38.347 [2024-10-29 10:58:43.830138] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:38.347 [2024-10-29 10:58:43.830372] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:38.347 [2024-10-29 10:58:43.830385] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:09:38.347 [2024-10-29 10:58:43.830543] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.347 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.608 [ 00:09:38.608 { 00:09:38.608 "name": "BaseBdev3", 00:09:38.608 "aliases": [ 00:09:38.608 "9e124678-5058-4291-a151-0215258f9ecf" 00:09:38.608 ], 00:09:38.608 "product_name": "Malloc disk", 00:09:38.608 "block_size": 512, 00:09:38.608 "num_blocks": 65536, 00:09:38.608 "uuid": "9e124678-5058-4291-a151-0215258f9ecf", 00:09:38.608 "assigned_rate_limits": { 00:09:38.608 "rw_ios_per_sec": 0, 00:09:38.608 "rw_mbytes_per_sec": 0, 00:09:38.608 "r_mbytes_per_sec": 0, 00:09:38.608 "w_mbytes_per_sec": 0 00:09:38.608 }, 00:09:38.608 "claimed": true, 00:09:38.608 "claim_type": "exclusive_write", 00:09:38.608 "zoned": false, 00:09:38.608 "supported_io_types": { 00:09:38.608 "read": true, 00:09:38.608 "write": true, 00:09:38.608 "unmap": true, 00:09:38.608 "flush": true, 00:09:38.608 "reset": true, 00:09:38.608 "nvme_admin": false, 00:09:38.608 "nvme_io": false, 00:09:38.608 "nvme_io_md": false, 00:09:38.608 "write_zeroes": true, 00:09:38.608 "zcopy": true, 00:09:38.608 "get_zone_info": false, 00:09:38.608 "zone_management": false, 00:09:38.608 "zone_append": false, 00:09:38.608 "compare": false, 00:09:38.608 "compare_and_write": false, 00:09:38.608 "abort": true, 00:09:38.608 "seek_hole": false, 00:09:38.608 "seek_data": false, 00:09:38.608 "copy": true, 00:09:38.608 "nvme_iov_md": false 00:09:38.608 }, 00:09:38.608 "memory_domains": [ 00:09:38.608 { 00:09:38.608 "dma_device_id": "system", 00:09:38.608 "dma_device_type": 1 00:09:38.608 }, 00:09:38.608 { 00:09:38.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.608 "dma_device_type": 2 00:09:38.608 } 00:09:38.608 ], 00:09:38.608 "driver_specific": {} 00:09:38.608 } 00:09:38.608 ] 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.608 "name": "Existed_Raid", 00:09:38.608 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:38.608 "strip_size_kb": 0, 00:09:38.608 "state": "online", 00:09:38.608 "raid_level": "raid1", 00:09:38.608 "superblock": true, 00:09:38.608 "num_base_bdevs": 3, 00:09:38.608 "num_base_bdevs_discovered": 3, 00:09:38.608 "num_base_bdevs_operational": 3, 00:09:38.608 "base_bdevs_list": [ 00:09:38.608 { 00:09:38.608 "name": "BaseBdev1", 00:09:38.608 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:38.608 "is_configured": true, 00:09:38.608 "data_offset": 2048, 00:09:38.608 "data_size": 63488 00:09:38.608 }, 00:09:38.608 { 00:09:38.608 "name": "BaseBdev2", 00:09:38.608 "uuid": "06b18932-65ed-4c40-bb45-d2f8635cdaad", 00:09:38.608 "is_configured": true, 00:09:38.608 "data_offset": 2048, 00:09:38.608 "data_size": 63488 00:09:38.608 }, 00:09:38.608 { 00:09:38.608 "name": "BaseBdev3", 00:09:38.608 "uuid": "9e124678-5058-4291-a151-0215258f9ecf", 00:09:38.608 "is_configured": true, 00:09:38.608 "data_offset": 2048, 00:09:38.608 "data_size": 63488 00:09:38.608 } 00:09:38.608 ] 00:09:38.608 }' 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.608 10:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:39.178 [2024-10-29 10:58:44.389009] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.178 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:39.178 "name": "Existed_Raid", 00:09:39.178 "aliases": [ 00:09:39.178 "1c4fc873-49bb-4399-a553-b67228539451" 00:09:39.178 ], 00:09:39.178 "product_name": "Raid Volume", 00:09:39.178 "block_size": 512, 00:09:39.178 "num_blocks": 63488, 00:09:39.178 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:39.178 "assigned_rate_limits": { 00:09:39.178 "rw_ios_per_sec": 0, 00:09:39.178 "rw_mbytes_per_sec": 0, 00:09:39.178 "r_mbytes_per_sec": 0, 00:09:39.178 "w_mbytes_per_sec": 0 00:09:39.178 }, 00:09:39.178 "claimed": false, 00:09:39.178 "zoned": false, 00:09:39.178 "supported_io_types": { 00:09:39.178 "read": true, 00:09:39.178 "write": true, 00:09:39.178 "unmap": false, 00:09:39.178 "flush": false, 00:09:39.178 "reset": true, 00:09:39.178 "nvme_admin": false, 00:09:39.178 "nvme_io": false, 00:09:39.178 "nvme_io_md": false, 00:09:39.178 "write_zeroes": true, 00:09:39.178 "zcopy": false, 00:09:39.178 "get_zone_info": false, 00:09:39.178 "zone_management": false, 00:09:39.178 "zone_append": false, 00:09:39.178 "compare": false, 00:09:39.178 "compare_and_write": false, 00:09:39.178 "abort": false, 00:09:39.178 "seek_hole": false, 00:09:39.178 "seek_data": false, 00:09:39.178 "copy": false, 00:09:39.178 "nvme_iov_md": false 00:09:39.178 }, 00:09:39.178 "memory_domains": [ 00:09:39.178 { 00:09:39.178 "dma_device_id": "system", 00:09:39.178 "dma_device_type": 1 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.178 "dma_device_type": 2 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "dma_device_id": "system", 00:09:39.178 "dma_device_type": 1 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.178 "dma_device_type": 2 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "dma_device_id": "system", 00:09:39.178 "dma_device_type": 1 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.178 "dma_device_type": 2 00:09:39.178 } 00:09:39.178 ], 00:09:39.178 "driver_specific": { 00:09:39.178 "raid": { 00:09:39.178 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:39.178 "strip_size_kb": 0, 00:09:39.178 "state": "online", 00:09:39.178 "raid_level": "raid1", 00:09:39.178 "superblock": true, 00:09:39.178 "num_base_bdevs": 3, 00:09:39.178 "num_base_bdevs_discovered": 3, 00:09:39.178 "num_base_bdevs_operational": 3, 00:09:39.178 "base_bdevs_list": [ 00:09:39.178 { 00:09:39.178 "name": "BaseBdev1", 00:09:39.178 "uuid": "937939e1-96bf-477a-a04f-08937cdd8faf", 00:09:39.178 "is_configured": true, 00:09:39.178 "data_offset": 2048, 00:09:39.178 "data_size": 63488 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "name": "BaseBdev2", 00:09:39.178 "uuid": "06b18932-65ed-4c40-bb45-d2f8635cdaad", 00:09:39.178 "is_configured": true, 00:09:39.178 "data_offset": 2048, 00:09:39.178 "data_size": 63488 00:09:39.178 }, 00:09:39.178 { 00:09:39.178 "name": "BaseBdev3", 00:09:39.178 "uuid": "9e124678-5058-4291-a151-0215258f9ecf", 00:09:39.178 "is_configured": true, 00:09:39.178 "data_offset": 2048, 00:09:39.179 "data_size": 63488 00:09:39.179 } 00:09:39.179 ] 00:09:39.179 } 00:09:39.179 } 00:09:39.179 }' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:39.179 BaseBdev2 00:09:39.179 BaseBdev3' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.179 [2024-10-29 10:58:44.644488] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.179 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.438 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.438 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.438 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.438 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.438 "name": "Existed_Raid", 00:09:39.438 "uuid": "1c4fc873-49bb-4399-a553-b67228539451", 00:09:39.438 "strip_size_kb": 0, 00:09:39.438 "state": "online", 00:09:39.438 "raid_level": "raid1", 00:09:39.438 "superblock": true, 00:09:39.438 "num_base_bdevs": 3, 00:09:39.438 "num_base_bdevs_discovered": 2, 00:09:39.438 "num_base_bdevs_operational": 2, 00:09:39.438 "base_bdevs_list": [ 00:09:39.438 { 00:09:39.438 "name": null, 00:09:39.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.438 "is_configured": false, 00:09:39.438 "data_offset": 0, 00:09:39.438 "data_size": 63488 00:09:39.438 }, 00:09:39.438 { 00:09:39.438 "name": "BaseBdev2", 00:09:39.438 "uuid": "06b18932-65ed-4c40-bb45-d2f8635cdaad", 00:09:39.439 "is_configured": true, 00:09:39.439 "data_offset": 2048, 00:09:39.439 "data_size": 63488 00:09:39.439 }, 00:09:39.439 { 00:09:39.439 "name": "BaseBdev3", 00:09:39.439 "uuid": "9e124678-5058-4291-a151-0215258f9ecf", 00:09:39.439 "is_configured": true, 00:09:39.439 "data_offset": 2048, 00:09:39.439 "data_size": 63488 00:09:39.439 } 00:09:39.439 ] 00:09:39.439 }' 00:09:39.439 10:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.439 10:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.697 [2024-10-29 10:58:45.148928] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:39.697 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.698 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.956 [2024-10-29 10:58:45.225552] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:39.956 [2024-10-29 10:58:45.225802] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:39.956 [2024-10-29 10:58:45.247952] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:39.956 [2024-10-29 10:58:45.248112] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:39.956 [2024-10-29 10:58:45.248170] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.956 BaseBdev2 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.956 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.956 [ 00:09:39.956 { 00:09:39.956 "name": "BaseBdev2", 00:09:39.956 "aliases": [ 00:09:39.956 "567984d2-43e1-4426-ae6b-f2bcbf865b44" 00:09:39.956 ], 00:09:39.956 "product_name": "Malloc disk", 00:09:39.956 "block_size": 512, 00:09:39.956 "num_blocks": 65536, 00:09:39.956 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:39.956 "assigned_rate_limits": { 00:09:39.956 "rw_ios_per_sec": 0, 00:09:39.956 "rw_mbytes_per_sec": 0, 00:09:39.956 "r_mbytes_per_sec": 0, 00:09:39.956 "w_mbytes_per_sec": 0 00:09:39.956 }, 00:09:39.956 "claimed": false, 00:09:39.956 "zoned": false, 00:09:39.956 "supported_io_types": { 00:09:39.956 "read": true, 00:09:39.956 "write": true, 00:09:39.956 "unmap": true, 00:09:39.956 "flush": true, 00:09:39.956 "reset": true, 00:09:39.956 "nvme_admin": false, 00:09:39.956 "nvme_io": false, 00:09:39.956 "nvme_io_md": false, 00:09:39.956 "write_zeroes": true, 00:09:39.956 "zcopy": true, 00:09:39.956 "get_zone_info": false, 00:09:39.956 "zone_management": false, 00:09:39.956 "zone_append": false, 00:09:39.956 "compare": false, 00:09:39.956 "compare_and_write": false, 00:09:39.956 "abort": true, 00:09:39.956 "seek_hole": false, 00:09:39.956 "seek_data": false, 00:09:39.956 "copy": true, 00:09:39.956 "nvme_iov_md": false 00:09:39.956 }, 00:09:39.956 "memory_domains": [ 00:09:39.956 { 00:09:39.956 "dma_device_id": "system", 00:09:39.956 "dma_device_type": 1 00:09:39.956 }, 00:09:39.956 { 00:09:39.956 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.956 "dma_device_type": 2 00:09:39.956 } 00:09:39.956 ], 00:09:39.957 "driver_specific": {} 00:09:39.957 } 00:09:39.957 ] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.957 BaseBdev3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.957 [ 00:09:39.957 { 00:09:39.957 "name": "BaseBdev3", 00:09:39.957 "aliases": [ 00:09:39.957 "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a" 00:09:39.957 ], 00:09:39.957 "product_name": "Malloc disk", 00:09:39.957 "block_size": 512, 00:09:39.957 "num_blocks": 65536, 00:09:39.957 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:39.957 "assigned_rate_limits": { 00:09:39.957 "rw_ios_per_sec": 0, 00:09:39.957 "rw_mbytes_per_sec": 0, 00:09:39.957 "r_mbytes_per_sec": 0, 00:09:39.957 "w_mbytes_per_sec": 0 00:09:39.957 }, 00:09:39.957 "claimed": false, 00:09:39.957 "zoned": false, 00:09:39.957 "supported_io_types": { 00:09:39.957 "read": true, 00:09:39.957 "write": true, 00:09:39.957 "unmap": true, 00:09:39.957 "flush": true, 00:09:39.957 "reset": true, 00:09:39.957 "nvme_admin": false, 00:09:39.957 "nvme_io": false, 00:09:39.957 "nvme_io_md": false, 00:09:39.957 "write_zeroes": true, 00:09:39.957 "zcopy": true, 00:09:39.957 "get_zone_info": false, 00:09:39.957 "zone_management": false, 00:09:39.957 "zone_append": false, 00:09:39.957 "compare": false, 00:09:39.957 "compare_and_write": false, 00:09:39.957 "abort": true, 00:09:39.957 "seek_hole": false, 00:09:39.957 "seek_data": false, 00:09:39.957 "copy": true, 00:09:39.957 "nvme_iov_md": false 00:09:39.957 }, 00:09:39.957 "memory_domains": [ 00:09:39.957 { 00:09:39.957 "dma_device_id": "system", 00:09:39.957 "dma_device_type": 1 00:09:39.957 }, 00:09:39.957 { 00:09:39.957 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.957 "dma_device_type": 2 00:09:39.957 } 00:09:39.957 ], 00:09:39.957 "driver_specific": {} 00:09:39.957 } 00:09:39.957 ] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.957 [2024-10-29 10:58:45.426428] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:39.957 [2024-10-29 10:58:45.426496] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:39.957 [2024-10-29 10:58:45.426527] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:39.957 [2024-10-29 10:58:45.428912] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.957 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.215 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.215 "name": "Existed_Raid", 00:09:40.215 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:40.215 "strip_size_kb": 0, 00:09:40.215 "state": "configuring", 00:09:40.215 "raid_level": "raid1", 00:09:40.215 "superblock": true, 00:09:40.215 "num_base_bdevs": 3, 00:09:40.215 "num_base_bdevs_discovered": 2, 00:09:40.215 "num_base_bdevs_operational": 3, 00:09:40.215 "base_bdevs_list": [ 00:09:40.215 { 00:09:40.215 "name": "BaseBdev1", 00:09:40.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.215 "is_configured": false, 00:09:40.215 "data_offset": 0, 00:09:40.215 "data_size": 0 00:09:40.215 }, 00:09:40.215 { 00:09:40.215 "name": "BaseBdev2", 00:09:40.215 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:40.215 "is_configured": true, 00:09:40.215 "data_offset": 2048, 00:09:40.215 "data_size": 63488 00:09:40.215 }, 00:09:40.215 { 00:09:40.215 "name": "BaseBdev3", 00:09:40.215 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:40.215 "is_configured": true, 00:09:40.215 "data_offset": 2048, 00:09:40.215 "data_size": 63488 00:09:40.215 } 00:09:40.215 ] 00:09:40.215 }' 00:09:40.215 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.215 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.474 [2024-10-29 10:58:45.837679] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.474 "name": "Existed_Raid", 00:09:40.474 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:40.474 "strip_size_kb": 0, 00:09:40.474 "state": "configuring", 00:09:40.474 "raid_level": "raid1", 00:09:40.474 "superblock": true, 00:09:40.474 "num_base_bdevs": 3, 00:09:40.474 "num_base_bdevs_discovered": 1, 00:09:40.474 "num_base_bdevs_operational": 3, 00:09:40.474 "base_bdevs_list": [ 00:09:40.474 { 00:09:40.474 "name": "BaseBdev1", 00:09:40.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.474 "is_configured": false, 00:09:40.474 "data_offset": 0, 00:09:40.474 "data_size": 0 00:09:40.474 }, 00:09:40.474 { 00:09:40.474 "name": null, 00:09:40.474 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:40.474 "is_configured": false, 00:09:40.474 "data_offset": 0, 00:09:40.474 "data_size": 63488 00:09:40.474 }, 00:09:40.474 { 00:09:40.474 "name": "BaseBdev3", 00:09:40.474 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:40.474 "is_configured": true, 00:09:40.474 "data_offset": 2048, 00:09:40.474 "data_size": 63488 00:09:40.474 } 00:09:40.474 ] 00:09:40.474 }' 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.474 10:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.038 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.038 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:41.038 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.038 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.038 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.039 [2024-10-29 10:58:46.297790] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:41.039 BaseBdev1 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.039 [ 00:09:41.039 { 00:09:41.039 "name": "BaseBdev1", 00:09:41.039 "aliases": [ 00:09:41.039 "457dbd04-12cb-4eca-be2e-3127f7011243" 00:09:41.039 ], 00:09:41.039 "product_name": "Malloc disk", 00:09:41.039 "block_size": 512, 00:09:41.039 "num_blocks": 65536, 00:09:41.039 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:41.039 "assigned_rate_limits": { 00:09:41.039 "rw_ios_per_sec": 0, 00:09:41.039 "rw_mbytes_per_sec": 0, 00:09:41.039 "r_mbytes_per_sec": 0, 00:09:41.039 "w_mbytes_per_sec": 0 00:09:41.039 }, 00:09:41.039 "claimed": true, 00:09:41.039 "claim_type": "exclusive_write", 00:09:41.039 "zoned": false, 00:09:41.039 "supported_io_types": { 00:09:41.039 "read": true, 00:09:41.039 "write": true, 00:09:41.039 "unmap": true, 00:09:41.039 "flush": true, 00:09:41.039 "reset": true, 00:09:41.039 "nvme_admin": false, 00:09:41.039 "nvme_io": false, 00:09:41.039 "nvme_io_md": false, 00:09:41.039 "write_zeroes": true, 00:09:41.039 "zcopy": true, 00:09:41.039 "get_zone_info": false, 00:09:41.039 "zone_management": false, 00:09:41.039 "zone_append": false, 00:09:41.039 "compare": false, 00:09:41.039 "compare_and_write": false, 00:09:41.039 "abort": true, 00:09:41.039 "seek_hole": false, 00:09:41.039 "seek_data": false, 00:09:41.039 "copy": true, 00:09:41.039 "nvme_iov_md": false 00:09:41.039 }, 00:09:41.039 "memory_domains": [ 00:09:41.039 { 00:09:41.039 "dma_device_id": "system", 00:09:41.039 "dma_device_type": 1 00:09:41.039 }, 00:09:41.039 { 00:09:41.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.039 "dma_device_type": 2 00:09:41.039 } 00:09:41.039 ], 00:09:41.039 "driver_specific": {} 00:09:41.039 } 00:09:41.039 ] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.039 "name": "Existed_Raid", 00:09:41.039 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:41.039 "strip_size_kb": 0, 00:09:41.039 "state": "configuring", 00:09:41.039 "raid_level": "raid1", 00:09:41.039 "superblock": true, 00:09:41.039 "num_base_bdevs": 3, 00:09:41.039 "num_base_bdevs_discovered": 2, 00:09:41.039 "num_base_bdevs_operational": 3, 00:09:41.039 "base_bdevs_list": [ 00:09:41.039 { 00:09:41.039 "name": "BaseBdev1", 00:09:41.039 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:41.039 "is_configured": true, 00:09:41.039 "data_offset": 2048, 00:09:41.039 "data_size": 63488 00:09:41.039 }, 00:09:41.039 { 00:09:41.039 "name": null, 00:09:41.039 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:41.039 "is_configured": false, 00:09:41.039 "data_offset": 0, 00:09:41.039 "data_size": 63488 00:09:41.039 }, 00:09:41.039 { 00:09:41.039 "name": "BaseBdev3", 00:09:41.039 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:41.039 "is_configured": true, 00:09:41.039 "data_offset": 2048, 00:09:41.039 "data_size": 63488 00:09:41.039 } 00:09:41.039 ] 00:09:41.039 }' 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.039 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.605 [2024-10-29 10:58:46.848978] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.605 "name": "Existed_Raid", 00:09:41.605 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:41.605 "strip_size_kb": 0, 00:09:41.605 "state": "configuring", 00:09:41.605 "raid_level": "raid1", 00:09:41.605 "superblock": true, 00:09:41.605 "num_base_bdevs": 3, 00:09:41.605 "num_base_bdevs_discovered": 1, 00:09:41.605 "num_base_bdevs_operational": 3, 00:09:41.605 "base_bdevs_list": [ 00:09:41.605 { 00:09:41.605 "name": "BaseBdev1", 00:09:41.605 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:41.605 "is_configured": true, 00:09:41.605 "data_offset": 2048, 00:09:41.605 "data_size": 63488 00:09:41.605 }, 00:09:41.605 { 00:09:41.605 "name": null, 00:09:41.605 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:41.605 "is_configured": false, 00:09:41.605 "data_offset": 0, 00:09:41.605 "data_size": 63488 00:09:41.605 }, 00:09:41.605 { 00:09:41.605 "name": null, 00:09:41.605 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:41.605 "is_configured": false, 00:09:41.605 "data_offset": 0, 00:09:41.605 "data_size": 63488 00:09:41.605 } 00:09:41.605 ] 00:09:41.605 }' 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.605 10:58:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.882 [2024-10-29 10:58:47.324175] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.882 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.176 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.176 "name": "Existed_Raid", 00:09:42.176 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:42.176 "strip_size_kb": 0, 00:09:42.176 "state": "configuring", 00:09:42.176 "raid_level": "raid1", 00:09:42.176 "superblock": true, 00:09:42.176 "num_base_bdevs": 3, 00:09:42.176 "num_base_bdevs_discovered": 2, 00:09:42.176 "num_base_bdevs_operational": 3, 00:09:42.176 "base_bdevs_list": [ 00:09:42.176 { 00:09:42.176 "name": "BaseBdev1", 00:09:42.176 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:42.176 "is_configured": true, 00:09:42.176 "data_offset": 2048, 00:09:42.176 "data_size": 63488 00:09:42.176 }, 00:09:42.176 { 00:09:42.176 "name": null, 00:09:42.176 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:42.176 "is_configured": false, 00:09:42.176 "data_offset": 0, 00:09:42.176 "data_size": 63488 00:09:42.176 }, 00:09:42.176 { 00:09:42.176 "name": "BaseBdev3", 00:09:42.176 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:42.176 "is_configured": true, 00:09:42.176 "data_offset": 2048, 00:09:42.176 "data_size": 63488 00:09:42.176 } 00:09:42.176 ] 00:09:42.176 }' 00:09:42.176 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.176 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.435 [2024-10-29 10:58:47.771584] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.435 "name": "Existed_Raid", 00:09:42.435 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:42.435 "strip_size_kb": 0, 00:09:42.435 "state": "configuring", 00:09:42.435 "raid_level": "raid1", 00:09:42.435 "superblock": true, 00:09:42.435 "num_base_bdevs": 3, 00:09:42.435 "num_base_bdevs_discovered": 1, 00:09:42.435 "num_base_bdevs_operational": 3, 00:09:42.435 "base_bdevs_list": [ 00:09:42.435 { 00:09:42.435 "name": null, 00:09:42.435 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:42.435 "is_configured": false, 00:09:42.435 "data_offset": 0, 00:09:42.435 "data_size": 63488 00:09:42.435 }, 00:09:42.435 { 00:09:42.435 "name": null, 00:09:42.435 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:42.435 "is_configured": false, 00:09:42.435 "data_offset": 0, 00:09:42.435 "data_size": 63488 00:09:42.435 }, 00:09:42.435 { 00:09:42.435 "name": "BaseBdev3", 00:09:42.435 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:42.435 "is_configured": true, 00:09:42.435 "data_offset": 2048, 00:09:42.435 "data_size": 63488 00:09:42.435 } 00:09:42.435 ] 00:09:42.435 }' 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.435 10:58:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.002 [2024-10-29 10:58:48.271045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.002 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.002 "name": "Existed_Raid", 00:09:43.002 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:43.002 "strip_size_kb": 0, 00:09:43.002 "state": "configuring", 00:09:43.002 "raid_level": "raid1", 00:09:43.002 "superblock": true, 00:09:43.002 "num_base_bdevs": 3, 00:09:43.002 "num_base_bdevs_discovered": 2, 00:09:43.002 "num_base_bdevs_operational": 3, 00:09:43.003 "base_bdevs_list": [ 00:09:43.003 { 00:09:43.003 "name": null, 00:09:43.003 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:43.003 "is_configured": false, 00:09:43.003 "data_offset": 0, 00:09:43.003 "data_size": 63488 00:09:43.003 }, 00:09:43.003 { 00:09:43.003 "name": "BaseBdev2", 00:09:43.003 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:43.003 "is_configured": true, 00:09:43.003 "data_offset": 2048, 00:09:43.003 "data_size": 63488 00:09:43.003 }, 00:09:43.003 { 00:09:43.003 "name": "BaseBdev3", 00:09:43.003 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:43.003 "is_configured": true, 00:09:43.003 "data_offset": 2048, 00:09:43.003 "data_size": 63488 00:09:43.003 } 00:09:43.003 ] 00:09:43.003 }' 00:09:43.003 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.003 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.261 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.262 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 457dbd04-12cb-4eca-be2e-3127f7011243 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.521 [2024-10-29 10:58:48.802924] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:43.521 [2024-10-29 10:58:48.803224] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:43.521 [2024-10-29 10:58:48.803241] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:43.521 [2024-10-29 10:58:48.803604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:09:43.521 [2024-10-29 10:58:48.803745] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:43.521 [2024-10-29 10:58:48.803762] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:09:43.521 NewBaseBdev 00:09:43.521 [2024-10-29 10:58:48.803900] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:43.521 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.522 [ 00:09:43.522 { 00:09:43.522 "name": "NewBaseBdev", 00:09:43.522 "aliases": [ 00:09:43.522 "457dbd04-12cb-4eca-be2e-3127f7011243" 00:09:43.522 ], 00:09:43.522 "product_name": "Malloc disk", 00:09:43.522 "block_size": 512, 00:09:43.522 "num_blocks": 65536, 00:09:43.522 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:43.522 "assigned_rate_limits": { 00:09:43.522 "rw_ios_per_sec": 0, 00:09:43.522 "rw_mbytes_per_sec": 0, 00:09:43.522 "r_mbytes_per_sec": 0, 00:09:43.522 "w_mbytes_per_sec": 0 00:09:43.522 }, 00:09:43.522 "claimed": true, 00:09:43.522 "claim_type": "exclusive_write", 00:09:43.522 "zoned": false, 00:09:43.522 "supported_io_types": { 00:09:43.522 "read": true, 00:09:43.522 "write": true, 00:09:43.522 "unmap": true, 00:09:43.522 "flush": true, 00:09:43.522 "reset": true, 00:09:43.522 "nvme_admin": false, 00:09:43.522 "nvme_io": false, 00:09:43.522 "nvme_io_md": false, 00:09:43.522 "write_zeroes": true, 00:09:43.522 "zcopy": true, 00:09:43.522 "get_zone_info": false, 00:09:43.522 "zone_management": false, 00:09:43.522 "zone_append": false, 00:09:43.522 "compare": false, 00:09:43.522 "compare_and_write": false, 00:09:43.522 "abort": true, 00:09:43.522 "seek_hole": false, 00:09:43.522 "seek_data": false, 00:09:43.522 "copy": true, 00:09:43.522 "nvme_iov_md": false 00:09:43.522 }, 00:09:43.522 "memory_domains": [ 00:09:43.522 { 00:09:43.522 "dma_device_id": "system", 00:09:43.522 "dma_device_type": 1 00:09:43.522 }, 00:09:43.522 { 00:09:43.522 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:43.522 "dma_device_type": 2 00:09:43.522 } 00:09:43.522 ], 00:09:43.522 "driver_specific": {} 00:09:43.522 } 00:09:43.522 ] 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.522 "name": "Existed_Raid", 00:09:43.522 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:43.522 "strip_size_kb": 0, 00:09:43.522 "state": "online", 00:09:43.522 "raid_level": "raid1", 00:09:43.522 "superblock": true, 00:09:43.522 "num_base_bdevs": 3, 00:09:43.522 "num_base_bdevs_discovered": 3, 00:09:43.522 "num_base_bdevs_operational": 3, 00:09:43.522 "base_bdevs_list": [ 00:09:43.522 { 00:09:43.522 "name": "NewBaseBdev", 00:09:43.522 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:43.522 "is_configured": true, 00:09:43.522 "data_offset": 2048, 00:09:43.522 "data_size": 63488 00:09:43.522 }, 00:09:43.522 { 00:09:43.522 "name": "BaseBdev2", 00:09:43.522 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:43.522 "is_configured": true, 00:09:43.522 "data_offset": 2048, 00:09:43.522 "data_size": 63488 00:09:43.522 }, 00:09:43.522 { 00:09:43.522 "name": "BaseBdev3", 00:09:43.522 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:43.522 "is_configured": true, 00:09:43.522 "data_offset": 2048, 00:09:43.522 "data_size": 63488 00:09:43.522 } 00:09:43.522 ] 00:09:43.522 }' 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.522 10:58:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:43.782 [2024-10-29 10:58:49.270566] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:43.782 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:44.042 "name": "Existed_Raid", 00:09:44.042 "aliases": [ 00:09:44.042 "02f5bc05-37dc-4415-b04b-6358869c0e0f" 00:09:44.042 ], 00:09:44.042 "product_name": "Raid Volume", 00:09:44.042 "block_size": 512, 00:09:44.042 "num_blocks": 63488, 00:09:44.042 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:44.042 "assigned_rate_limits": { 00:09:44.042 "rw_ios_per_sec": 0, 00:09:44.042 "rw_mbytes_per_sec": 0, 00:09:44.042 "r_mbytes_per_sec": 0, 00:09:44.042 "w_mbytes_per_sec": 0 00:09:44.042 }, 00:09:44.042 "claimed": false, 00:09:44.042 "zoned": false, 00:09:44.042 "supported_io_types": { 00:09:44.042 "read": true, 00:09:44.042 "write": true, 00:09:44.042 "unmap": false, 00:09:44.042 "flush": false, 00:09:44.042 "reset": true, 00:09:44.042 "nvme_admin": false, 00:09:44.042 "nvme_io": false, 00:09:44.042 "nvme_io_md": false, 00:09:44.042 "write_zeroes": true, 00:09:44.042 "zcopy": false, 00:09:44.042 "get_zone_info": false, 00:09:44.042 "zone_management": false, 00:09:44.042 "zone_append": false, 00:09:44.042 "compare": false, 00:09:44.042 "compare_and_write": false, 00:09:44.042 "abort": false, 00:09:44.042 "seek_hole": false, 00:09:44.042 "seek_data": false, 00:09:44.042 "copy": false, 00:09:44.042 "nvme_iov_md": false 00:09:44.042 }, 00:09:44.042 "memory_domains": [ 00:09:44.042 { 00:09:44.042 "dma_device_id": "system", 00:09:44.042 "dma_device_type": 1 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.042 "dma_device_type": 2 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "dma_device_id": "system", 00:09:44.042 "dma_device_type": 1 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.042 "dma_device_type": 2 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "dma_device_id": "system", 00:09:44.042 "dma_device_type": 1 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.042 "dma_device_type": 2 00:09:44.042 } 00:09:44.042 ], 00:09:44.042 "driver_specific": { 00:09:44.042 "raid": { 00:09:44.042 "uuid": "02f5bc05-37dc-4415-b04b-6358869c0e0f", 00:09:44.042 "strip_size_kb": 0, 00:09:44.042 "state": "online", 00:09:44.042 "raid_level": "raid1", 00:09:44.042 "superblock": true, 00:09:44.042 "num_base_bdevs": 3, 00:09:44.042 "num_base_bdevs_discovered": 3, 00:09:44.042 "num_base_bdevs_operational": 3, 00:09:44.042 "base_bdevs_list": [ 00:09:44.042 { 00:09:44.042 "name": "NewBaseBdev", 00:09:44.042 "uuid": "457dbd04-12cb-4eca-be2e-3127f7011243", 00:09:44.042 "is_configured": true, 00:09:44.042 "data_offset": 2048, 00:09:44.042 "data_size": 63488 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "name": "BaseBdev2", 00:09:44.042 "uuid": "567984d2-43e1-4426-ae6b-f2bcbf865b44", 00:09:44.042 "is_configured": true, 00:09:44.042 "data_offset": 2048, 00:09:44.042 "data_size": 63488 00:09:44.042 }, 00:09:44.042 { 00:09:44.042 "name": "BaseBdev3", 00:09:44.042 "uuid": "f81ae2d3-ab77-4f57-a2bb-78fafcffcb2a", 00:09:44.042 "is_configured": true, 00:09:44.042 "data_offset": 2048, 00:09:44.042 "data_size": 63488 00:09:44.042 } 00:09:44.042 ] 00:09:44.042 } 00:09:44.042 } 00:09:44.042 }' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:44.042 BaseBdev2 00:09:44.042 BaseBdev3' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.042 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.043 [2024-10-29 10:58:49.517748] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:44.043 [2024-10-29 10:58:49.517794] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:44.043 [2024-10-29 10:58:49.517885] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:44.043 [2024-10-29 10:58:49.518164] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:44.043 [2024-10-29 10:58:49.518174] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 79203 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 79203 ']' 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 79203 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:44.043 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 79203 00:09:44.307 killing process with pid 79203 00:09:44.307 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:44.307 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:44.307 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 79203' 00:09:44.307 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 79203 00:09:44.307 [2024-10-29 10:58:49.566128] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:44.307 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 79203 00:09:44.307 [2024-10-29 10:58:49.627545] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:44.566 ************************************ 00:09:44.566 END TEST raid_state_function_test_sb 00:09:44.566 ************************************ 00:09:44.566 10:58:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:44.566 00:09:44.566 real 0m8.902s 00:09:44.566 user 0m14.915s 00:09:44.566 sys 0m1.852s 00:09:44.566 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:44.566 10:58:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:44.566 10:58:50 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:44.566 10:58:50 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:09:44.566 10:58:50 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:44.566 10:58:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:44.566 ************************************ 00:09:44.566 START TEST raid_superblock_test 00:09:44.566 ************************************ 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 3 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79808 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79808 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 79808 ']' 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:44.566 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:44.566 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.825 [2024-10-29 10:58:50.112314] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:44.825 [2024-10-29 10:58:50.112913] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79808 ] 00:09:44.825 [2024-10-29 10:58:50.305350] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:45.085 [2024-10-29 10:58:50.347932] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:45.085 [2024-10-29 10:58:50.425338] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:45.085 [2024-10-29 10:58:50.425431] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 malloc1 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 [2024-10-29 10:58:50.976163] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:45.655 [2024-10-29 10:58:50.976369] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.655 [2024-10-29 10:58:50.976439] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:45.655 [2024-10-29 10:58:50.976496] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.655 [2024-10-29 10:58:50.979335] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.655 [2024-10-29 10:58:50.979456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:45.655 pt1 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 malloc2 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 [2024-10-29 10:58:51.014959] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:45.655 [2024-10-29 10:58:51.015081] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.655 [2024-10-29 10:58:51.015102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:45.655 [2024-10-29 10:58:51.015114] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.655 [2024-10-29 10:58:51.017589] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.655 [2024-10-29 10:58:51.017626] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:45.655 pt2 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 malloc3 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 [2024-10-29 10:58:51.049910] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:45.655 [2024-10-29 10:58:51.050041] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.655 [2024-10-29 10:58:51.050080] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:45.655 [2024-10-29 10:58:51.050111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.655 [2024-10-29 10:58:51.052598] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.655 [2024-10-29 10:58:51.052671] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:45.655 pt3 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.655 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.655 [2024-10-29 10:58:51.061929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:45.655 [2024-10-29 10:58:51.064170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:45.655 [2024-10-29 10:58:51.064268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:45.655 [2024-10-29 10:58:51.064469] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:09:45.655 [2024-10-29 10:58:51.064519] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:45.655 [2024-10-29 10:58:51.064873] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:45.655 [2024-10-29 10:58:51.065086] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:09:45.656 [2024-10-29 10:58:51.065133] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:09:45.656 [2024-10-29 10:58:51.065312] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.656 "name": "raid_bdev1", 00:09:45.656 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:45.656 "strip_size_kb": 0, 00:09:45.656 "state": "online", 00:09:45.656 "raid_level": "raid1", 00:09:45.656 "superblock": true, 00:09:45.656 "num_base_bdevs": 3, 00:09:45.656 "num_base_bdevs_discovered": 3, 00:09:45.656 "num_base_bdevs_operational": 3, 00:09:45.656 "base_bdevs_list": [ 00:09:45.656 { 00:09:45.656 "name": "pt1", 00:09:45.656 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:45.656 "is_configured": true, 00:09:45.656 "data_offset": 2048, 00:09:45.656 "data_size": 63488 00:09:45.656 }, 00:09:45.656 { 00:09:45.656 "name": "pt2", 00:09:45.656 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:45.656 "is_configured": true, 00:09:45.656 "data_offset": 2048, 00:09:45.656 "data_size": 63488 00:09:45.656 }, 00:09:45.656 { 00:09:45.656 "name": "pt3", 00:09:45.656 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:45.656 "is_configured": true, 00:09:45.656 "data_offset": 2048, 00:09:45.656 "data_size": 63488 00:09:45.656 } 00:09:45.656 ] 00:09:45.656 }' 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.656 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.226 [2024-10-29 10:58:51.493555] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:46.226 "name": "raid_bdev1", 00:09:46.226 "aliases": [ 00:09:46.226 "c9f4ad8a-eded-43c7-904f-413eecf209cd" 00:09:46.226 ], 00:09:46.226 "product_name": "Raid Volume", 00:09:46.226 "block_size": 512, 00:09:46.226 "num_blocks": 63488, 00:09:46.226 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:46.226 "assigned_rate_limits": { 00:09:46.226 "rw_ios_per_sec": 0, 00:09:46.226 "rw_mbytes_per_sec": 0, 00:09:46.226 "r_mbytes_per_sec": 0, 00:09:46.226 "w_mbytes_per_sec": 0 00:09:46.226 }, 00:09:46.226 "claimed": false, 00:09:46.226 "zoned": false, 00:09:46.226 "supported_io_types": { 00:09:46.226 "read": true, 00:09:46.226 "write": true, 00:09:46.226 "unmap": false, 00:09:46.226 "flush": false, 00:09:46.226 "reset": true, 00:09:46.226 "nvme_admin": false, 00:09:46.226 "nvme_io": false, 00:09:46.226 "nvme_io_md": false, 00:09:46.226 "write_zeroes": true, 00:09:46.226 "zcopy": false, 00:09:46.226 "get_zone_info": false, 00:09:46.226 "zone_management": false, 00:09:46.226 "zone_append": false, 00:09:46.226 "compare": false, 00:09:46.226 "compare_and_write": false, 00:09:46.226 "abort": false, 00:09:46.226 "seek_hole": false, 00:09:46.226 "seek_data": false, 00:09:46.226 "copy": false, 00:09:46.226 "nvme_iov_md": false 00:09:46.226 }, 00:09:46.226 "memory_domains": [ 00:09:46.226 { 00:09:46.226 "dma_device_id": "system", 00:09:46.226 "dma_device_type": 1 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.226 "dma_device_type": 2 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "dma_device_id": "system", 00:09:46.226 "dma_device_type": 1 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.226 "dma_device_type": 2 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "dma_device_id": "system", 00:09:46.226 "dma_device_type": 1 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.226 "dma_device_type": 2 00:09:46.226 } 00:09:46.226 ], 00:09:46.226 "driver_specific": { 00:09:46.226 "raid": { 00:09:46.226 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:46.226 "strip_size_kb": 0, 00:09:46.226 "state": "online", 00:09:46.226 "raid_level": "raid1", 00:09:46.226 "superblock": true, 00:09:46.226 "num_base_bdevs": 3, 00:09:46.226 "num_base_bdevs_discovered": 3, 00:09:46.226 "num_base_bdevs_operational": 3, 00:09:46.226 "base_bdevs_list": [ 00:09:46.226 { 00:09:46.226 "name": "pt1", 00:09:46.226 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:46.226 "is_configured": true, 00:09:46.226 "data_offset": 2048, 00:09:46.226 "data_size": 63488 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "name": "pt2", 00:09:46.226 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:46.226 "is_configured": true, 00:09:46.226 "data_offset": 2048, 00:09:46.226 "data_size": 63488 00:09:46.226 }, 00:09:46.226 { 00:09:46.226 "name": "pt3", 00:09:46.226 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:46.226 "is_configured": true, 00:09:46.226 "data_offset": 2048, 00:09:46.226 "data_size": 63488 00:09:46.226 } 00:09:46.226 ] 00:09:46.226 } 00:09:46.226 } 00:09:46.226 }' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:46.226 pt2 00:09:46.226 pt3' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.226 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.486 [2024-10-29 10:58:51.780997] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c9f4ad8a-eded-43c7-904f-413eecf209cd 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z c9f4ad8a-eded-43c7-904f-413eecf209cd ']' 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.486 [2024-10-29 10:58:51.812628] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:46.486 [2024-10-29 10:58:51.812665] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:46.486 [2024-10-29 10:58:51.812777] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:46.486 [2024-10-29 10:58:51.812879] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:46.486 [2024-10-29 10:58:51.812896] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.486 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 [2024-10-29 10:58:51.964400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:46.487 [2024-10-29 10:58:51.966818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:46.487 [2024-10-29 10:58:51.966914] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:46.487 [2024-10-29 10:58:51.967002] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:46.487 [2024-10-29 10:58:51.967103] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:46.487 [2024-10-29 10:58:51.967127] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:46.487 [2024-10-29 10:58:51.967142] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:46.487 [2024-10-29 10:58:51.967155] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:09:46.487 request: 00:09:46.487 { 00:09:46.487 "name": "raid_bdev1", 00:09:46.487 "raid_level": "raid1", 00:09:46.487 "base_bdevs": [ 00:09:46.487 "malloc1", 00:09:46.487 "malloc2", 00:09:46.487 "malloc3" 00:09:46.487 ], 00:09:46.487 "superblock": false, 00:09:46.487 "method": "bdev_raid_create", 00:09:46.487 "req_id": 1 00:09:46.487 } 00:09:46.487 Got JSON-RPC error response 00:09:46.487 response: 00:09:46.487 { 00:09:46.487 "code": -17, 00:09:46.487 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:46.487 } 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.487 10:58:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:46.747 10:58:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.747 [2024-10-29 10:58:52.032217] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:46.747 [2024-10-29 10:58:52.032348] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:46.747 [2024-10-29 10:58:52.032422] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:46.747 [2024-10-29 10:58:52.032455] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:46.747 [2024-10-29 10:58:52.035016] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:46.747 [2024-10-29 10:58:52.035092] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:46.747 [2024-10-29 10:58:52.035214] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:46.747 [2024-10-29 10:58:52.035280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:46.747 pt1 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.747 "name": "raid_bdev1", 00:09:46.747 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:46.747 "strip_size_kb": 0, 00:09:46.747 "state": "configuring", 00:09:46.747 "raid_level": "raid1", 00:09:46.747 "superblock": true, 00:09:46.747 "num_base_bdevs": 3, 00:09:46.747 "num_base_bdevs_discovered": 1, 00:09:46.747 "num_base_bdevs_operational": 3, 00:09:46.747 "base_bdevs_list": [ 00:09:46.747 { 00:09:46.747 "name": "pt1", 00:09:46.747 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:46.747 "is_configured": true, 00:09:46.747 "data_offset": 2048, 00:09:46.747 "data_size": 63488 00:09:46.747 }, 00:09:46.747 { 00:09:46.747 "name": null, 00:09:46.747 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:46.747 "is_configured": false, 00:09:46.747 "data_offset": 2048, 00:09:46.747 "data_size": 63488 00:09:46.747 }, 00:09:46.747 { 00:09:46.747 "name": null, 00:09:46.747 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:46.747 "is_configured": false, 00:09:46.747 "data_offset": 2048, 00:09:46.747 "data_size": 63488 00:09:46.747 } 00:09:46.747 ] 00:09:46.747 }' 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.747 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.007 [2024-10-29 10:58:52.439577] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:47.007 [2024-10-29 10:58:52.439754] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.007 [2024-10-29 10:58:52.439786] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:47.007 [2024-10-29 10:58:52.439803] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.007 [2024-10-29 10:58:52.440376] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.007 [2024-10-29 10:58:52.440401] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:47.007 [2024-10-29 10:58:52.440511] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:47.007 [2024-10-29 10:58:52.440543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:47.007 pt2 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.007 [2024-10-29 10:58:52.451561] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.007 "name": "raid_bdev1", 00:09:47.007 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:47.007 "strip_size_kb": 0, 00:09:47.007 "state": "configuring", 00:09:47.007 "raid_level": "raid1", 00:09:47.007 "superblock": true, 00:09:47.007 "num_base_bdevs": 3, 00:09:47.007 "num_base_bdevs_discovered": 1, 00:09:47.007 "num_base_bdevs_operational": 3, 00:09:47.007 "base_bdevs_list": [ 00:09:47.007 { 00:09:47.007 "name": "pt1", 00:09:47.007 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:47.007 "is_configured": true, 00:09:47.007 "data_offset": 2048, 00:09:47.007 "data_size": 63488 00:09:47.007 }, 00:09:47.007 { 00:09:47.007 "name": null, 00:09:47.007 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:47.007 "is_configured": false, 00:09:47.007 "data_offset": 0, 00:09:47.007 "data_size": 63488 00:09:47.007 }, 00:09:47.007 { 00:09:47.007 "name": null, 00:09:47.007 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:47.007 "is_configured": false, 00:09:47.007 "data_offset": 2048, 00:09:47.007 "data_size": 63488 00:09:47.007 } 00:09:47.007 ] 00:09:47.007 }' 00:09:47.007 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.267 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.528 [2024-10-29 10:58:52.834891] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:47.528 [2024-10-29 10:58:52.835043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.528 [2024-10-29 10:58:52.835087] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:47.528 [2024-10-29 10:58:52.835117] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.528 [2024-10-29 10:58:52.835715] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.528 [2024-10-29 10:58:52.835786] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:47.528 [2024-10-29 10:58:52.835931] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:47.528 [2024-10-29 10:58:52.835994] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:47.528 pt2 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.528 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.528 [2024-10-29 10:58:52.846816] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:47.528 [2024-10-29 10:58:52.846908] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.528 [2024-10-29 10:58:52.846948] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:47.528 [2024-10-29 10:58:52.846977] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.528 [2024-10-29 10:58:52.847479] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.528 [2024-10-29 10:58:52.847545] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:47.528 [2024-10-29 10:58:52.847674] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:47.528 [2024-10-29 10:58:52.847777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:47.529 [2024-10-29 10:58:52.847937] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:09:47.529 [2024-10-29 10:58:52.847975] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:47.529 [2024-10-29 10:58:52.848288] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:09:47.529 [2024-10-29 10:58:52.848485] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:09:47.529 [2024-10-29 10:58:52.848545] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:09:47.529 [2024-10-29 10:58:52.848674] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.529 pt3 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.529 "name": "raid_bdev1", 00:09:47.529 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:47.529 "strip_size_kb": 0, 00:09:47.529 "state": "online", 00:09:47.529 "raid_level": "raid1", 00:09:47.529 "superblock": true, 00:09:47.529 "num_base_bdevs": 3, 00:09:47.529 "num_base_bdevs_discovered": 3, 00:09:47.529 "num_base_bdevs_operational": 3, 00:09:47.529 "base_bdevs_list": [ 00:09:47.529 { 00:09:47.529 "name": "pt1", 00:09:47.529 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:47.529 "is_configured": true, 00:09:47.529 "data_offset": 2048, 00:09:47.529 "data_size": 63488 00:09:47.529 }, 00:09:47.529 { 00:09:47.529 "name": "pt2", 00:09:47.529 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:47.529 "is_configured": true, 00:09:47.529 "data_offset": 2048, 00:09:47.529 "data_size": 63488 00:09:47.529 }, 00:09:47.529 { 00:09:47.529 "name": "pt3", 00:09:47.529 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:47.529 "is_configured": true, 00:09:47.529 "data_offset": 2048, 00:09:47.529 "data_size": 63488 00:09:47.529 } 00:09:47.529 ] 00:09:47.529 }' 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.529 10:58:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.789 [2024-10-29 10:58:53.234546] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.789 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:47.789 "name": "raid_bdev1", 00:09:47.789 "aliases": [ 00:09:47.789 "c9f4ad8a-eded-43c7-904f-413eecf209cd" 00:09:47.789 ], 00:09:47.789 "product_name": "Raid Volume", 00:09:47.789 "block_size": 512, 00:09:47.789 "num_blocks": 63488, 00:09:47.789 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:47.789 "assigned_rate_limits": { 00:09:47.789 "rw_ios_per_sec": 0, 00:09:47.789 "rw_mbytes_per_sec": 0, 00:09:47.789 "r_mbytes_per_sec": 0, 00:09:47.789 "w_mbytes_per_sec": 0 00:09:47.789 }, 00:09:47.789 "claimed": false, 00:09:47.789 "zoned": false, 00:09:47.789 "supported_io_types": { 00:09:47.789 "read": true, 00:09:47.789 "write": true, 00:09:47.789 "unmap": false, 00:09:47.789 "flush": false, 00:09:47.789 "reset": true, 00:09:47.789 "nvme_admin": false, 00:09:47.789 "nvme_io": false, 00:09:47.789 "nvme_io_md": false, 00:09:47.789 "write_zeroes": true, 00:09:47.789 "zcopy": false, 00:09:47.789 "get_zone_info": false, 00:09:47.789 "zone_management": false, 00:09:47.789 "zone_append": false, 00:09:47.789 "compare": false, 00:09:47.789 "compare_and_write": false, 00:09:47.789 "abort": false, 00:09:47.789 "seek_hole": false, 00:09:47.789 "seek_data": false, 00:09:47.789 "copy": false, 00:09:47.789 "nvme_iov_md": false 00:09:47.789 }, 00:09:47.789 "memory_domains": [ 00:09:47.789 { 00:09:47.789 "dma_device_id": "system", 00:09:47.789 "dma_device_type": 1 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.789 "dma_device_type": 2 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "dma_device_id": "system", 00:09:47.789 "dma_device_type": 1 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.789 "dma_device_type": 2 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "dma_device_id": "system", 00:09:47.789 "dma_device_type": 1 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.789 "dma_device_type": 2 00:09:47.789 } 00:09:47.789 ], 00:09:47.789 "driver_specific": { 00:09:47.789 "raid": { 00:09:47.789 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:47.789 "strip_size_kb": 0, 00:09:47.789 "state": "online", 00:09:47.789 "raid_level": "raid1", 00:09:47.789 "superblock": true, 00:09:47.789 "num_base_bdevs": 3, 00:09:47.789 "num_base_bdevs_discovered": 3, 00:09:47.789 "num_base_bdevs_operational": 3, 00:09:47.789 "base_bdevs_list": [ 00:09:47.789 { 00:09:47.789 "name": "pt1", 00:09:47.789 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:47.789 "is_configured": true, 00:09:47.789 "data_offset": 2048, 00:09:47.789 "data_size": 63488 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "name": "pt2", 00:09:47.789 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:47.789 "is_configured": true, 00:09:47.789 "data_offset": 2048, 00:09:47.789 "data_size": 63488 00:09:47.789 }, 00:09:47.789 { 00:09:47.789 "name": "pt3", 00:09:47.789 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:47.789 "is_configured": true, 00:09:47.789 "data_offset": 2048, 00:09:47.789 "data_size": 63488 00:09:47.789 } 00:09:47.789 ] 00:09:47.789 } 00:09:47.789 } 00:09:47.789 }' 00:09:47.790 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:48.049 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:48.049 pt2 00:09:48.049 pt3' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.050 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.050 [2024-10-29 10:58:53.529996] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' c9f4ad8a-eded-43c7-904f-413eecf209cd '!=' c9f4ad8a-eded-43c7-904f-413eecf209cd ']' 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.311 [2024-10-29 10:58:53.573674] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.311 "name": "raid_bdev1", 00:09:48.311 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:48.311 "strip_size_kb": 0, 00:09:48.311 "state": "online", 00:09:48.311 "raid_level": "raid1", 00:09:48.311 "superblock": true, 00:09:48.311 "num_base_bdevs": 3, 00:09:48.311 "num_base_bdevs_discovered": 2, 00:09:48.311 "num_base_bdevs_operational": 2, 00:09:48.311 "base_bdevs_list": [ 00:09:48.311 { 00:09:48.311 "name": null, 00:09:48.311 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.311 "is_configured": false, 00:09:48.311 "data_offset": 0, 00:09:48.311 "data_size": 63488 00:09:48.311 }, 00:09:48.311 { 00:09:48.311 "name": "pt2", 00:09:48.311 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.311 "is_configured": true, 00:09:48.311 "data_offset": 2048, 00:09:48.311 "data_size": 63488 00:09:48.311 }, 00:09:48.311 { 00:09:48.311 "name": "pt3", 00:09:48.311 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.311 "is_configured": true, 00:09:48.311 "data_offset": 2048, 00:09:48.311 "data_size": 63488 00:09:48.311 } 00:09:48.311 ] 00:09:48.311 }' 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.311 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.571 10:58:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:48.571 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.571 10:58:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.571 [2024-10-29 10:58:53.996958] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:48.571 [2024-10-29 10:58:53.997076] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:48.571 [2024-10-29 10:58:53.997199] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:48.571 [2024-10-29 10:58:53.997324] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:48.571 [2024-10-29 10:58:53.997401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.571 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.831 [2024-10-29 10:58:54.080757] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:48.831 [2024-10-29 10:58:54.080822] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.831 [2024-10-29 10:58:54.080844] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:48.831 [2024-10-29 10:58:54.080854] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.831 [2024-10-29 10:58:54.083469] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.831 [2024-10-29 10:58:54.083557] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:48.831 [2024-10-29 10:58:54.083646] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:48.831 [2024-10-29 10:58:54.083681] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:48.831 pt2 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.831 "name": "raid_bdev1", 00:09:48.831 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:48.831 "strip_size_kb": 0, 00:09:48.831 "state": "configuring", 00:09:48.831 "raid_level": "raid1", 00:09:48.831 "superblock": true, 00:09:48.831 "num_base_bdevs": 3, 00:09:48.831 "num_base_bdevs_discovered": 1, 00:09:48.831 "num_base_bdevs_operational": 2, 00:09:48.831 "base_bdevs_list": [ 00:09:48.831 { 00:09:48.831 "name": null, 00:09:48.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.831 "is_configured": false, 00:09:48.831 "data_offset": 2048, 00:09:48.831 "data_size": 63488 00:09:48.831 }, 00:09:48.831 { 00:09:48.831 "name": "pt2", 00:09:48.831 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.831 "is_configured": true, 00:09:48.831 "data_offset": 2048, 00:09:48.831 "data_size": 63488 00:09:48.831 }, 00:09:48.831 { 00:09:48.831 "name": null, 00:09:48.831 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.831 "is_configured": false, 00:09:48.831 "data_offset": 2048, 00:09:48.831 "data_size": 63488 00:09:48.831 } 00:09:48.831 ] 00:09:48.831 }' 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.831 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.091 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.091 [2024-10-29 10:58:54.516149] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:49.091 [2024-10-29 10:58:54.516328] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:49.091 [2024-10-29 10:58:54.516386] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:09:49.091 [2024-10-29 10:58:54.516433] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:49.091 [2024-10-29 10:58:54.517001] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:49.091 [2024-10-29 10:58:54.517069] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:49.091 [2024-10-29 10:58:54.517223] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:49.091 [2024-10-29 10:58:54.517285] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:49.091 [2024-10-29 10:58:54.517462] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:49.091 [2024-10-29 10:58:54.517499] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:49.091 [2024-10-29 10:58:54.517814] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:49.091 [2024-10-29 10:58:54.517995] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:49.091 [2024-10-29 10:58:54.518037] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:49.091 [2024-10-29 10:58:54.518208] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:49.091 pt3 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.092 "name": "raid_bdev1", 00:09:49.092 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:49.092 "strip_size_kb": 0, 00:09:49.092 "state": "online", 00:09:49.092 "raid_level": "raid1", 00:09:49.092 "superblock": true, 00:09:49.092 "num_base_bdevs": 3, 00:09:49.092 "num_base_bdevs_discovered": 2, 00:09:49.092 "num_base_bdevs_operational": 2, 00:09:49.092 "base_bdevs_list": [ 00:09:49.092 { 00:09:49.092 "name": null, 00:09:49.092 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.092 "is_configured": false, 00:09:49.092 "data_offset": 2048, 00:09:49.092 "data_size": 63488 00:09:49.092 }, 00:09:49.092 { 00:09:49.092 "name": "pt2", 00:09:49.092 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.092 "is_configured": true, 00:09:49.092 "data_offset": 2048, 00:09:49.092 "data_size": 63488 00:09:49.092 }, 00:09:49.092 { 00:09:49.092 "name": "pt3", 00:09:49.092 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.092 "is_configured": true, 00:09:49.092 "data_offset": 2048, 00:09:49.092 "data_size": 63488 00:09:49.092 } 00:09:49.092 ] 00:09:49.092 }' 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.092 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 [2024-10-29 10:58:54.927473] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:49.662 [2024-10-29 10:58:54.927518] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:49.662 [2024-10-29 10:58:54.927628] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:49.662 [2024-10-29 10:58:54.927697] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:49.662 [2024-10-29 10:58:54.927711] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.662 10:58:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 [2024-10-29 10:58:55.003284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:49.662 [2024-10-29 10:58:55.003387] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:49.662 [2024-10-29 10:58:55.003409] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:09:49.662 [2024-10-29 10:58:55.003422] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:49.662 [2024-10-29 10:58:55.006154] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:49.662 [2024-10-29 10:58:55.006195] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:49.662 [2024-10-29 10:58:55.006284] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:49.662 [2024-10-29 10:58:55.006332] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:49.662 [2024-10-29 10:58:55.006467] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:49.662 [2024-10-29 10:58:55.006491] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:49.662 [2024-10-29 10:58:55.006508] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:09:49.662 [2024-10-29 10:58:55.006551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:49.662 pt1 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.662 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.662 "name": "raid_bdev1", 00:09:49.662 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:49.662 "strip_size_kb": 0, 00:09:49.662 "state": "configuring", 00:09:49.662 "raid_level": "raid1", 00:09:49.662 "superblock": true, 00:09:49.662 "num_base_bdevs": 3, 00:09:49.662 "num_base_bdevs_discovered": 1, 00:09:49.662 "num_base_bdevs_operational": 2, 00:09:49.662 "base_bdevs_list": [ 00:09:49.662 { 00:09:49.662 "name": null, 00:09:49.662 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.662 "is_configured": false, 00:09:49.662 "data_offset": 2048, 00:09:49.662 "data_size": 63488 00:09:49.662 }, 00:09:49.662 { 00:09:49.662 "name": "pt2", 00:09:49.662 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.662 "is_configured": true, 00:09:49.662 "data_offset": 2048, 00:09:49.662 "data_size": 63488 00:09:49.662 }, 00:09:49.663 { 00:09:49.663 "name": null, 00:09:49.663 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.663 "is_configured": false, 00:09:49.663 "data_offset": 2048, 00:09:49.663 "data_size": 63488 00:09:49.663 } 00:09:49.663 ] 00:09:49.663 }' 00:09:49.663 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.663 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.233 [2024-10-29 10:58:55.478523] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:50.233 [2024-10-29 10:58:55.478698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:50.233 [2024-10-29 10:58:55.478744] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:09:50.233 [2024-10-29 10:58:55.478795] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:50.233 [2024-10-29 10:58:55.479428] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:50.233 [2024-10-29 10:58:55.479507] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:50.233 [2024-10-29 10:58:55.479655] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:50.233 [2024-10-29 10:58:55.479756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:50.233 [2024-10-29 10:58:55.479930] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:09:50.233 [2024-10-29 10:58:55.479977] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:50.233 [2024-10-29 10:58:55.480292] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:09:50.233 [2024-10-29 10:58:55.480595] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:09:50.233 [2024-10-29 10:58:55.480655] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:09:50.233 [2024-10-29 10:58:55.480848] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:50.233 pt3 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.233 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.234 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.234 "name": "raid_bdev1", 00:09:50.234 "uuid": "c9f4ad8a-eded-43c7-904f-413eecf209cd", 00:09:50.234 "strip_size_kb": 0, 00:09:50.234 "state": "online", 00:09:50.234 "raid_level": "raid1", 00:09:50.234 "superblock": true, 00:09:50.234 "num_base_bdevs": 3, 00:09:50.234 "num_base_bdevs_discovered": 2, 00:09:50.234 "num_base_bdevs_operational": 2, 00:09:50.234 "base_bdevs_list": [ 00:09:50.234 { 00:09:50.234 "name": null, 00:09:50.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.234 "is_configured": false, 00:09:50.234 "data_offset": 2048, 00:09:50.234 "data_size": 63488 00:09:50.234 }, 00:09:50.234 { 00:09:50.234 "name": "pt2", 00:09:50.234 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:50.234 "is_configured": true, 00:09:50.234 "data_offset": 2048, 00:09:50.234 "data_size": 63488 00:09:50.234 }, 00:09:50.234 { 00:09:50.234 "name": "pt3", 00:09:50.234 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:50.234 "is_configured": true, 00:09:50.234 "data_offset": 2048, 00:09:50.234 "data_size": 63488 00:09:50.234 } 00:09:50.234 ] 00:09:50.234 }' 00:09:50.234 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.234 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.506 [2024-10-29 10:58:55.957996] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' c9f4ad8a-eded-43c7-904f-413eecf209cd '!=' c9f4ad8a-eded-43c7-904f-413eecf209cd ']' 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79808 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 79808 ']' 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 79808 00:09:50.506 10:58:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 79808 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:50.775 killing process with pid 79808 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 79808' 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 79808 00:09:50.775 [2024-10-29 10:58:56.038152] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:50.775 [2024-10-29 10:58:56.038275] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:50.775 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 79808 00:09:50.775 [2024-10-29 10:58:56.038348] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:50.775 [2024-10-29 10:58:56.038375] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:09:50.776 [2024-10-29 10:58:56.102242] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:51.035 10:58:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:51.036 00:09:51.036 real 0m6.404s 00:09:51.036 user 0m10.503s 00:09:51.036 sys 0m1.403s 00:09:51.036 ************************************ 00:09:51.036 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:51.036 10:58:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.036 END TEST raid_superblock_test 00:09:51.036 ************************************ 00:09:51.036 10:58:56 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:51.036 10:58:56 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:51.036 10:58:56 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:51.036 10:58:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:51.036 ************************************ 00:09:51.036 START TEST raid_read_error_test 00:09:51.036 ************************************ 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 3 read 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.mtKV0hRS7h 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80237 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80237 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 80237 ']' 00:09:51.036 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:51.036 10:58:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.295 [2024-10-29 10:58:56.606284] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:51.295 [2024-10-29 10:58:56.606426] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80237 ] 00:09:51.295 [2024-10-29 10:58:56.776177] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:51.555 [2024-10-29 10:58:56.819892] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:51.555 [2024-10-29 10:58:56.898425] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:51.555 [2024-10-29 10:58:56.898473] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.125 BaseBdev1_malloc 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.125 true 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.125 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 [2024-10-29 10:58:57.485918] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:52.126 [2024-10-29 10:58:57.485991] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.126 [2024-10-29 10:58:57.486018] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:52.126 [2024-10-29 10:58:57.486035] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.126 [2024-10-29 10:58:57.488519] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.126 [2024-10-29 10:58:57.488643] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:52.126 BaseBdev1 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 BaseBdev2_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 true 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 [2024-10-29 10:58:57.532686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:52.126 [2024-10-29 10:58:57.532751] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.126 [2024-10-29 10:58:57.532772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:52.126 [2024-10-29 10:58:57.532780] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.126 [2024-10-29 10:58:57.535169] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.126 [2024-10-29 10:58:57.535288] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:52.126 BaseBdev2 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 BaseBdev3_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 true 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 [2024-10-29 10:58:57.579389] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:52.126 [2024-10-29 10:58:57.579449] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:52.126 [2024-10-29 10:58:57.579471] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:52.126 [2024-10-29 10:58:57.579481] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:52.126 [2024-10-29 10:58:57.581913] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:52.126 [2024-10-29 10:58:57.582018] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:52.126 BaseBdev3 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 [2024-10-29 10:58:57.591417] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:52.126 [2024-10-29 10:58:57.593592] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:52.126 [2024-10-29 10:58:57.593681] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:52.126 [2024-10-29 10:58:57.593872] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:52.126 [2024-10-29 10:58:57.593887] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:52.126 [2024-10-29 10:58:57.594138] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:52.126 [2024-10-29 10:58:57.594304] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:52.126 [2024-10-29 10:58:57.594315] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:52.126 [2024-10-29 10:58:57.594467] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.126 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.386 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.386 "name": "raid_bdev1", 00:09:52.386 "uuid": "adb6ba24-2a24-414a-9d8d-84479512f706", 00:09:52.386 "strip_size_kb": 0, 00:09:52.386 "state": "online", 00:09:52.386 "raid_level": "raid1", 00:09:52.386 "superblock": true, 00:09:52.387 "num_base_bdevs": 3, 00:09:52.387 "num_base_bdevs_discovered": 3, 00:09:52.387 "num_base_bdevs_operational": 3, 00:09:52.387 "base_bdevs_list": [ 00:09:52.387 { 00:09:52.387 "name": "BaseBdev1", 00:09:52.387 "uuid": "b3ec4181-a999-5e96-9e4e-af3a085d5a62", 00:09:52.387 "is_configured": true, 00:09:52.387 "data_offset": 2048, 00:09:52.387 "data_size": 63488 00:09:52.387 }, 00:09:52.387 { 00:09:52.387 "name": "BaseBdev2", 00:09:52.387 "uuid": "9145221e-ef64-5be2-9767-ac332da60ea0", 00:09:52.387 "is_configured": true, 00:09:52.387 "data_offset": 2048, 00:09:52.387 "data_size": 63488 00:09:52.387 }, 00:09:52.387 { 00:09:52.387 "name": "BaseBdev3", 00:09:52.387 "uuid": "04a9cf7e-4551-5075-9bda-d6ba8e343437", 00:09:52.387 "is_configured": true, 00:09:52.387 "data_offset": 2048, 00:09:52.387 "data_size": 63488 00:09:52.387 } 00:09:52.387 ] 00:09:52.387 }' 00:09:52.387 10:58:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.387 10:58:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.647 10:58:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:52.647 10:58:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:52.647 [2024-10-29 10:58:58.122996] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.586 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.846 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.846 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.846 "name": "raid_bdev1", 00:09:53.846 "uuid": "adb6ba24-2a24-414a-9d8d-84479512f706", 00:09:53.846 "strip_size_kb": 0, 00:09:53.846 "state": "online", 00:09:53.846 "raid_level": "raid1", 00:09:53.846 "superblock": true, 00:09:53.846 "num_base_bdevs": 3, 00:09:53.846 "num_base_bdevs_discovered": 3, 00:09:53.846 "num_base_bdevs_operational": 3, 00:09:53.846 "base_bdevs_list": [ 00:09:53.846 { 00:09:53.846 "name": "BaseBdev1", 00:09:53.846 "uuid": "b3ec4181-a999-5e96-9e4e-af3a085d5a62", 00:09:53.846 "is_configured": true, 00:09:53.846 "data_offset": 2048, 00:09:53.846 "data_size": 63488 00:09:53.846 }, 00:09:53.846 { 00:09:53.846 "name": "BaseBdev2", 00:09:53.846 "uuid": "9145221e-ef64-5be2-9767-ac332da60ea0", 00:09:53.846 "is_configured": true, 00:09:53.846 "data_offset": 2048, 00:09:53.846 "data_size": 63488 00:09:53.846 }, 00:09:53.846 { 00:09:53.846 "name": "BaseBdev3", 00:09:53.846 "uuid": "04a9cf7e-4551-5075-9bda-d6ba8e343437", 00:09:53.846 "is_configured": true, 00:09:53.846 "data_offset": 2048, 00:09:53.846 "data_size": 63488 00:09:53.846 } 00:09:53.846 ] 00:09:53.846 }' 00:09:53.846 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.846 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.106 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:54.106 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.106 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.106 [2024-10-29 10:58:59.472607] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:54.106 [2024-10-29 10:58:59.472774] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:54.106 [2024-10-29 10:58:59.475441] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:54.106 [2024-10-29 10:58:59.475499] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:54.106 [2024-10-29 10:58:59.475609] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:54.106 [2024-10-29 10:58:59.475622] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:54.106 { 00:09:54.106 "results": [ 00:09:54.106 { 00:09:54.106 "job": "raid_bdev1", 00:09:54.106 "core_mask": "0x1", 00:09:54.106 "workload": "randrw", 00:09:54.106 "percentage": 50, 00:09:54.106 "status": "finished", 00:09:54.106 "queue_depth": 1, 00:09:54.106 "io_size": 131072, 00:09:54.106 "runtime": 1.350106, 00:09:54.106 "iops": 10525.099510705086, 00:09:54.106 "mibps": 1315.6374388381357, 00:09:54.106 "io_failed": 0, 00:09:54.106 "io_timeout": 0, 00:09:54.106 "avg_latency_us": 92.35901600754742, 00:09:54.107 "min_latency_us": 23.475982532751093, 00:09:54.107 "max_latency_us": 2675.814847161572 00:09:54.107 } 00:09:54.107 ], 00:09:54.107 "core_count": 1 00:09:54.107 } 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80237 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 80237 ']' 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 80237 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 80237 00:09:54.107 killing process with pid 80237 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 80237' 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 80237 00:09:54.107 [2024-10-29 10:58:59.524593] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:54.107 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 80237 00:09:54.107 [2024-10-29 10:58:59.574326] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.mtKV0hRS7h 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:54.675 00:09:54.675 real 0m3.393s 00:09:54.675 user 0m4.197s 00:09:54.675 sys 0m0.585s 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:54.675 10:58:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.675 ************************************ 00:09:54.675 END TEST raid_read_error_test 00:09:54.675 ************************************ 00:09:54.675 10:58:59 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:54.675 10:58:59 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:54.675 10:58:59 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:54.675 10:58:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:54.675 ************************************ 00:09:54.675 START TEST raid_write_error_test 00:09:54.675 ************************************ 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 3 write 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.34ebSlM41Y 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80366 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80366 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 80366 ']' 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:54.675 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:54.675 10:58:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.675 [2024-10-29 10:59:00.072260] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:54.675 [2024-10-29 10:59:00.072839] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80366 ] 00:09:54.934 [2024-10-29 10:59:00.240914] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:54.934 [2024-10-29 10:59:00.283392] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:54.934 [2024-10-29 10:59:00.361038] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:54.934 [2024-10-29 10:59:00.361145] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 BaseBdev1_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 true 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 [2024-10-29 10:59:00.952975] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:55.502 [2024-10-29 10:59:00.953117] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.502 [2024-10-29 10:59:00.953152] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:09:55.502 [2024-10-29 10:59:00.953161] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.502 [2024-10-29 10:59:00.955648] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.502 [2024-10-29 10:59:00.955692] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:55.502 BaseBdev1 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 BaseBdev2_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 true 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.502 10:59:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.502 [2024-10-29 10:59:01.000357] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:55.502 [2024-10-29 10:59:01.000431] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.502 [2024-10-29 10:59:01.000451] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:09:55.502 [2024-10-29 10:59:01.000462] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.763 [2024-10-29 10:59:01.002891] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.763 [2024-10-29 10:59:01.002927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:55.763 BaseBdev2 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.763 BaseBdev3_malloc 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.763 true 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.763 [2024-10-29 10:59:01.047094] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:55.763 [2024-10-29 10:59:01.047227] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:55.763 [2024-10-29 10:59:01.047254] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:55.763 [2024-10-29 10:59:01.047264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:55.763 [2024-10-29 10:59:01.049741] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:55.763 [2024-10-29 10:59:01.049779] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:55.763 BaseBdev3 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.763 [2024-10-29 10:59:01.059134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:55.763 [2024-10-29 10:59:01.061299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:55.763 [2024-10-29 10:59:01.061407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:55.763 [2024-10-29 10:59:01.061617] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:09:55.763 [2024-10-29 10:59:01.061637] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:55.763 [2024-10-29 10:59:01.061911] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:09:55.763 [2024-10-29 10:59:01.062108] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:09:55.763 [2024-10-29 10:59:01.062119] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:09:55.763 [2024-10-29 10:59:01.062290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.763 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.763 "name": "raid_bdev1", 00:09:55.763 "uuid": "7f6170d2-ace4-4d5b-b97a-31e2418ee325", 00:09:55.763 "strip_size_kb": 0, 00:09:55.763 "state": "online", 00:09:55.763 "raid_level": "raid1", 00:09:55.763 "superblock": true, 00:09:55.764 "num_base_bdevs": 3, 00:09:55.764 "num_base_bdevs_discovered": 3, 00:09:55.764 "num_base_bdevs_operational": 3, 00:09:55.764 "base_bdevs_list": [ 00:09:55.764 { 00:09:55.764 "name": "BaseBdev1", 00:09:55.764 "uuid": "a3eee654-ed10-5292-97a2-ddf1068ae071", 00:09:55.764 "is_configured": true, 00:09:55.764 "data_offset": 2048, 00:09:55.764 "data_size": 63488 00:09:55.764 }, 00:09:55.764 { 00:09:55.764 "name": "BaseBdev2", 00:09:55.764 "uuid": "124d3c73-39e6-5b84-9384-8b0b36a9ba86", 00:09:55.764 "is_configured": true, 00:09:55.764 "data_offset": 2048, 00:09:55.764 "data_size": 63488 00:09:55.764 }, 00:09:55.764 { 00:09:55.764 "name": "BaseBdev3", 00:09:55.764 "uuid": "dc92fd0f-5fb2-5d3b-9382-91b1689b3343", 00:09:55.764 "is_configured": true, 00:09:55.764 "data_offset": 2048, 00:09:55.764 "data_size": 63488 00:09:55.764 } 00:09:55.764 ] 00:09:55.764 }' 00:09:55.764 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.764 10:59:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.334 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:56.334 10:59:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:56.334 [2024-10-29 10:59:01.614729] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.272 [2024-10-29 10:59:02.535122] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:57.272 [2024-10-29 10:59:02.535280] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:57.272 [2024-10-29 10:59:02.535591] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005e10 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.272 "name": "raid_bdev1", 00:09:57.272 "uuid": "7f6170d2-ace4-4d5b-b97a-31e2418ee325", 00:09:57.272 "strip_size_kb": 0, 00:09:57.272 "state": "online", 00:09:57.272 "raid_level": "raid1", 00:09:57.272 "superblock": true, 00:09:57.272 "num_base_bdevs": 3, 00:09:57.272 "num_base_bdevs_discovered": 2, 00:09:57.272 "num_base_bdevs_operational": 2, 00:09:57.272 "base_bdevs_list": [ 00:09:57.272 { 00:09:57.272 "name": null, 00:09:57.272 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.272 "is_configured": false, 00:09:57.272 "data_offset": 0, 00:09:57.272 "data_size": 63488 00:09:57.272 }, 00:09:57.272 { 00:09:57.272 "name": "BaseBdev2", 00:09:57.272 "uuid": "124d3c73-39e6-5b84-9384-8b0b36a9ba86", 00:09:57.272 "is_configured": true, 00:09:57.272 "data_offset": 2048, 00:09:57.272 "data_size": 63488 00:09:57.272 }, 00:09:57.272 { 00:09:57.272 "name": "BaseBdev3", 00:09:57.272 "uuid": "dc92fd0f-5fb2-5d3b-9382-91b1689b3343", 00:09:57.272 "is_configured": true, 00:09:57.272 "data_offset": 2048, 00:09:57.272 "data_size": 63488 00:09:57.272 } 00:09:57.272 ] 00:09:57.272 }' 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.272 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.533 10:59:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:57.533 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.533 10:59:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.533 [2024-10-29 10:59:02.999875] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:57.533 [2024-10-29 10:59:03.000012] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.533 [2024-10-29 10:59:03.002532] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.533 [2024-10-29 10:59:03.002581] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:57.533 [2024-10-29 10:59:03.002671] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:57.533 [2024-10-29 10:59:03.002681] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:09:57.533 { 00:09:57.533 "results": [ 00:09:57.533 { 00:09:57.533 "job": "raid_bdev1", 00:09:57.533 "core_mask": "0x1", 00:09:57.533 "workload": "randrw", 00:09:57.533 "percentage": 50, 00:09:57.533 "status": "finished", 00:09:57.533 "queue_depth": 1, 00:09:57.533 "io_size": 131072, 00:09:57.533 "runtime": 1.385637, 00:09:57.533 "iops": 11958.398916888045, 00:09:57.533 "mibps": 1494.7998646110057, 00:09:57.533 "io_failed": 0, 00:09:57.533 "io_timeout": 0, 00:09:57.533 "avg_latency_us": 81.01546594703427, 00:09:57.533 "min_latency_us": 23.252401746724892, 00:09:57.533 "max_latency_us": 1373.6803493449781 00:09:57.533 } 00:09:57.533 ], 00:09:57.533 "core_count": 1 00:09:57.533 } 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80366 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 80366 ']' 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 80366 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:09:57.533 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 80366 00:09:57.794 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:09:57.794 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:09:57.794 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 80366' 00:09:57.794 killing process with pid 80366 00:09:57.794 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 80366 00:09:57.794 [2024-10-29 10:59:03.047030] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:57.794 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 80366 00:09:57.794 [2024-10-29 10:59:03.095028] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.34ebSlM41Y 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:58.055 00:09:58.055 real 0m3.457s 00:09:58.055 user 0m4.287s 00:09:58.055 sys 0m0.606s 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:09:58.055 10:59:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.055 ************************************ 00:09:58.055 END TEST raid_write_error_test 00:09:58.055 ************************************ 00:09:58.055 10:59:03 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:58.055 10:59:03 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:58.055 10:59:03 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:58.055 10:59:03 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:09:58.055 10:59:03 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:09:58.055 10:59:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:58.055 ************************************ 00:09:58.055 START TEST raid_state_function_test 00:09:58.055 ************************************ 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 4 false 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80499 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80499' 00:09:58.055 Process raid pid: 80499 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80499 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 80499 ']' 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:58.055 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:09:58.055 10:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.316 [2024-10-29 10:59:03.604058] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:09:58.316 [2024-10-29 10:59:03.604291] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:58.316 [2024-10-29 10:59:03.774871] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:58.579 [2024-10-29 10:59:03.819296] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:58.579 [2024-10-29 10:59:03.896666] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:58.579 [2024-10-29 10:59:03.896838] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.157 [2024-10-29 10:59:04.433416] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:59.157 [2024-10-29 10:59:04.433573] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:59.157 [2024-10-29 10:59:04.433603] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:59.157 [2024-10-29 10:59:04.433630] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:59.157 [2024-10-29 10:59:04.433648] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:59.157 [2024-10-29 10:59:04.433672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:59.157 [2024-10-29 10:59:04.433689] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:59.157 [2024-10-29 10:59:04.433738] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.157 "name": "Existed_Raid", 00:09:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.157 "strip_size_kb": 64, 00:09:59.157 "state": "configuring", 00:09:59.157 "raid_level": "raid0", 00:09:59.157 "superblock": false, 00:09:59.157 "num_base_bdevs": 4, 00:09:59.157 "num_base_bdevs_discovered": 0, 00:09:59.157 "num_base_bdevs_operational": 4, 00:09:59.157 "base_bdevs_list": [ 00:09:59.157 { 00:09:59.157 "name": "BaseBdev1", 00:09:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.157 "is_configured": false, 00:09:59.157 "data_offset": 0, 00:09:59.157 "data_size": 0 00:09:59.157 }, 00:09:59.157 { 00:09:59.157 "name": "BaseBdev2", 00:09:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.157 "is_configured": false, 00:09:59.157 "data_offset": 0, 00:09:59.157 "data_size": 0 00:09:59.157 }, 00:09:59.157 { 00:09:59.157 "name": "BaseBdev3", 00:09:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.157 "is_configured": false, 00:09:59.157 "data_offset": 0, 00:09:59.157 "data_size": 0 00:09:59.157 }, 00:09:59.157 { 00:09:59.157 "name": "BaseBdev4", 00:09:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.157 "is_configured": false, 00:09:59.157 "data_offset": 0, 00:09:59.157 "data_size": 0 00:09:59.157 } 00:09:59.157 ] 00:09:59.157 }' 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.157 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.434 [2024-10-29 10:59:04.884591] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:59.434 [2024-10-29 10:59:04.884748] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.434 [2024-10-29 10:59:04.896530] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:59.434 [2024-10-29 10:59:04.896617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:59.434 [2024-10-29 10:59:04.896643] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:59.434 [2024-10-29 10:59:04.896665] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:59.434 [2024-10-29 10:59:04.896683] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:59.434 [2024-10-29 10:59:04.896703] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:59.434 [2024-10-29 10:59:04.896720] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:59.434 [2024-10-29 10:59:04.896740] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.434 [2024-10-29 10:59:04.923910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:59.434 BaseBdev1 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.434 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.705 [ 00:09:59.705 { 00:09:59.705 "name": "BaseBdev1", 00:09:59.705 "aliases": [ 00:09:59.705 "c8edfbd8-6636-4b5d-b18f-9401d481ff53" 00:09:59.705 ], 00:09:59.705 "product_name": "Malloc disk", 00:09:59.705 "block_size": 512, 00:09:59.705 "num_blocks": 65536, 00:09:59.705 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:09:59.705 "assigned_rate_limits": { 00:09:59.705 "rw_ios_per_sec": 0, 00:09:59.705 "rw_mbytes_per_sec": 0, 00:09:59.705 "r_mbytes_per_sec": 0, 00:09:59.705 "w_mbytes_per_sec": 0 00:09:59.705 }, 00:09:59.705 "claimed": true, 00:09:59.705 "claim_type": "exclusive_write", 00:09:59.705 "zoned": false, 00:09:59.705 "supported_io_types": { 00:09:59.705 "read": true, 00:09:59.705 "write": true, 00:09:59.705 "unmap": true, 00:09:59.705 "flush": true, 00:09:59.705 "reset": true, 00:09:59.705 "nvme_admin": false, 00:09:59.705 "nvme_io": false, 00:09:59.705 "nvme_io_md": false, 00:09:59.705 "write_zeroes": true, 00:09:59.705 "zcopy": true, 00:09:59.705 "get_zone_info": false, 00:09:59.705 "zone_management": false, 00:09:59.705 "zone_append": false, 00:09:59.705 "compare": false, 00:09:59.705 "compare_and_write": false, 00:09:59.705 "abort": true, 00:09:59.705 "seek_hole": false, 00:09:59.705 "seek_data": false, 00:09:59.705 "copy": true, 00:09:59.705 "nvme_iov_md": false 00:09:59.705 }, 00:09:59.705 "memory_domains": [ 00:09:59.705 { 00:09:59.705 "dma_device_id": "system", 00:09:59.705 "dma_device_type": 1 00:09:59.705 }, 00:09:59.705 { 00:09:59.705 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.705 "dma_device_type": 2 00:09:59.705 } 00:09:59.705 ], 00:09:59.705 "driver_specific": {} 00:09:59.705 } 00:09:59.705 ] 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.705 10:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.705 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.705 "name": "Existed_Raid", 00:09:59.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.705 "strip_size_kb": 64, 00:09:59.705 "state": "configuring", 00:09:59.705 "raid_level": "raid0", 00:09:59.705 "superblock": false, 00:09:59.705 "num_base_bdevs": 4, 00:09:59.705 "num_base_bdevs_discovered": 1, 00:09:59.705 "num_base_bdevs_operational": 4, 00:09:59.705 "base_bdevs_list": [ 00:09:59.705 { 00:09:59.705 "name": "BaseBdev1", 00:09:59.705 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:09:59.705 "is_configured": true, 00:09:59.705 "data_offset": 0, 00:09:59.705 "data_size": 65536 00:09:59.705 }, 00:09:59.705 { 00:09:59.705 "name": "BaseBdev2", 00:09:59.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.705 "is_configured": false, 00:09:59.705 "data_offset": 0, 00:09:59.705 "data_size": 0 00:09:59.705 }, 00:09:59.705 { 00:09:59.705 "name": "BaseBdev3", 00:09:59.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.705 "is_configured": false, 00:09:59.705 "data_offset": 0, 00:09:59.705 "data_size": 0 00:09:59.705 }, 00:09:59.705 { 00:09:59.705 "name": "BaseBdev4", 00:09:59.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.705 "is_configured": false, 00:09:59.705 "data_offset": 0, 00:09:59.705 "data_size": 0 00:09:59.705 } 00:09:59.705 ] 00:09:59.705 }' 00:09:59.705 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.705 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.965 [2024-10-29 10:59:05.391289] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:59.965 [2024-10-29 10:59:05.391409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.965 [2024-10-29 10:59:05.403270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:59.965 [2024-10-29 10:59:05.405843] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:59.965 [2024-10-29 10:59:05.405894] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:59.965 [2024-10-29 10:59:05.405905] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:59.965 [2024-10-29 10:59:05.405916] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:59.965 [2024-10-29 10:59:05.405923] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:59.965 [2024-10-29 10:59:05.405933] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.965 "name": "Existed_Raid", 00:09:59.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.965 "strip_size_kb": 64, 00:09:59.965 "state": "configuring", 00:09:59.965 "raid_level": "raid0", 00:09:59.965 "superblock": false, 00:09:59.965 "num_base_bdevs": 4, 00:09:59.965 "num_base_bdevs_discovered": 1, 00:09:59.965 "num_base_bdevs_operational": 4, 00:09:59.965 "base_bdevs_list": [ 00:09:59.965 { 00:09:59.965 "name": "BaseBdev1", 00:09:59.965 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:09:59.965 "is_configured": true, 00:09:59.965 "data_offset": 0, 00:09:59.965 "data_size": 65536 00:09:59.965 }, 00:09:59.965 { 00:09:59.965 "name": "BaseBdev2", 00:09:59.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.965 "is_configured": false, 00:09:59.965 "data_offset": 0, 00:09:59.965 "data_size": 0 00:09:59.965 }, 00:09:59.965 { 00:09:59.965 "name": "BaseBdev3", 00:09:59.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.965 "is_configured": false, 00:09:59.965 "data_offset": 0, 00:09:59.965 "data_size": 0 00:09:59.965 }, 00:09:59.965 { 00:09:59.965 "name": "BaseBdev4", 00:09:59.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.965 "is_configured": false, 00:09:59.965 "data_offset": 0, 00:09:59.965 "data_size": 0 00:09:59.965 } 00:09:59.965 ] 00:09:59.965 }' 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.965 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.534 [2024-10-29 10:59:05.879943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:00.534 BaseBdev2 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.534 [ 00:10:00.534 { 00:10:00.534 "name": "BaseBdev2", 00:10:00.534 "aliases": [ 00:10:00.534 "6a49e1f6-47ff-48b6-99fc-1a1353404588" 00:10:00.534 ], 00:10:00.534 "product_name": "Malloc disk", 00:10:00.534 "block_size": 512, 00:10:00.534 "num_blocks": 65536, 00:10:00.534 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:00.534 "assigned_rate_limits": { 00:10:00.534 "rw_ios_per_sec": 0, 00:10:00.534 "rw_mbytes_per_sec": 0, 00:10:00.534 "r_mbytes_per_sec": 0, 00:10:00.534 "w_mbytes_per_sec": 0 00:10:00.534 }, 00:10:00.534 "claimed": true, 00:10:00.534 "claim_type": "exclusive_write", 00:10:00.534 "zoned": false, 00:10:00.534 "supported_io_types": { 00:10:00.534 "read": true, 00:10:00.534 "write": true, 00:10:00.534 "unmap": true, 00:10:00.534 "flush": true, 00:10:00.534 "reset": true, 00:10:00.534 "nvme_admin": false, 00:10:00.534 "nvme_io": false, 00:10:00.534 "nvme_io_md": false, 00:10:00.534 "write_zeroes": true, 00:10:00.534 "zcopy": true, 00:10:00.534 "get_zone_info": false, 00:10:00.534 "zone_management": false, 00:10:00.534 "zone_append": false, 00:10:00.534 "compare": false, 00:10:00.534 "compare_and_write": false, 00:10:00.534 "abort": true, 00:10:00.534 "seek_hole": false, 00:10:00.534 "seek_data": false, 00:10:00.534 "copy": true, 00:10:00.534 "nvme_iov_md": false 00:10:00.534 }, 00:10:00.534 "memory_domains": [ 00:10:00.534 { 00:10:00.534 "dma_device_id": "system", 00:10:00.534 "dma_device_type": 1 00:10:00.534 }, 00:10:00.534 { 00:10:00.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.534 "dma_device_type": 2 00:10:00.534 } 00:10:00.534 ], 00:10:00.534 "driver_specific": {} 00:10:00.534 } 00:10:00.534 ] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.534 "name": "Existed_Raid", 00:10:00.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.534 "strip_size_kb": 64, 00:10:00.534 "state": "configuring", 00:10:00.534 "raid_level": "raid0", 00:10:00.534 "superblock": false, 00:10:00.534 "num_base_bdevs": 4, 00:10:00.534 "num_base_bdevs_discovered": 2, 00:10:00.534 "num_base_bdevs_operational": 4, 00:10:00.534 "base_bdevs_list": [ 00:10:00.534 { 00:10:00.534 "name": "BaseBdev1", 00:10:00.534 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:10:00.534 "is_configured": true, 00:10:00.534 "data_offset": 0, 00:10:00.534 "data_size": 65536 00:10:00.534 }, 00:10:00.534 { 00:10:00.534 "name": "BaseBdev2", 00:10:00.534 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:00.534 "is_configured": true, 00:10:00.534 "data_offset": 0, 00:10:00.534 "data_size": 65536 00:10:00.534 }, 00:10:00.534 { 00:10:00.534 "name": "BaseBdev3", 00:10:00.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.534 "is_configured": false, 00:10:00.534 "data_offset": 0, 00:10:00.534 "data_size": 0 00:10:00.534 }, 00:10:00.534 { 00:10:00.534 "name": "BaseBdev4", 00:10:00.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.534 "is_configured": false, 00:10:00.534 "data_offset": 0, 00:10:00.534 "data_size": 0 00:10:00.534 } 00:10:00.534 ] 00:10:00.534 }' 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.534 10:59:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.103 [2024-10-29 10:59:06.368748] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:01.103 BaseBdev3 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.103 [ 00:10:01.103 { 00:10:01.103 "name": "BaseBdev3", 00:10:01.103 "aliases": [ 00:10:01.103 "b4d4cf77-2359-41b0-8b46-963dd05d8b6e" 00:10:01.103 ], 00:10:01.103 "product_name": "Malloc disk", 00:10:01.103 "block_size": 512, 00:10:01.103 "num_blocks": 65536, 00:10:01.103 "uuid": "b4d4cf77-2359-41b0-8b46-963dd05d8b6e", 00:10:01.103 "assigned_rate_limits": { 00:10:01.103 "rw_ios_per_sec": 0, 00:10:01.103 "rw_mbytes_per_sec": 0, 00:10:01.103 "r_mbytes_per_sec": 0, 00:10:01.103 "w_mbytes_per_sec": 0 00:10:01.103 }, 00:10:01.103 "claimed": true, 00:10:01.103 "claim_type": "exclusive_write", 00:10:01.103 "zoned": false, 00:10:01.103 "supported_io_types": { 00:10:01.103 "read": true, 00:10:01.103 "write": true, 00:10:01.103 "unmap": true, 00:10:01.103 "flush": true, 00:10:01.103 "reset": true, 00:10:01.103 "nvme_admin": false, 00:10:01.103 "nvme_io": false, 00:10:01.103 "nvme_io_md": false, 00:10:01.103 "write_zeroes": true, 00:10:01.103 "zcopy": true, 00:10:01.103 "get_zone_info": false, 00:10:01.103 "zone_management": false, 00:10:01.103 "zone_append": false, 00:10:01.103 "compare": false, 00:10:01.103 "compare_and_write": false, 00:10:01.103 "abort": true, 00:10:01.103 "seek_hole": false, 00:10:01.103 "seek_data": false, 00:10:01.103 "copy": true, 00:10:01.103 "nvme_iov_md": false 00:10:01.103 }, 00:10:01.103 "memory_domains": [ 00:10:01.103 { 00:10:01.103 "dma_device_id": "system", 00:10:01.103 "dma_device_type": 1 00:10:01.103 }, 00:10:01.103 { 00:10:01.103 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.103 "dma_device_type": 2 00:10:01.103 } 00:10:01.103 ], 00:10:01.103 "driver_specific": {} 00:10:01.103 } 00:10:01.103 ] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.103 "name": "Existed_Raid", 00:10:01.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.103 "strip_size_kb": 64, 00:10:01.103 "state": "configuring", 00:10:01.103 "raid_level": "raid0", 00:10:01.103 "superblock": false, 00:10:01.103 "num_base_bdevs": 4, 00:10:01.103 "num_base_bdevs_discovered": 3, 00:10:01.103 "num_base_bdevs_operational": 4, 00:10:01.103 "base_bdevs_list": [ 00:10:01.103 { 00:10:01.103 "name": "BaseBdev1", 00:10:01.103 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:10:01.103 "is_configured": true, 00:10:01.103 "data_offset": 0, 00:10:01.103 "data_size": 65536 00:10:01.103 }, 00:10:01.103 { 00:10:01.103 "name": "BaseBdev2", 00:10:01.103 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:01.103 "is_configured": true, 00:10:01.103 "data_offset": 0, 00:10:01.103 "data_size": 65536 00:10:01.103 }, 00:10:01.103 { 00:10:01.103 "name": "BaseBdev3", 00:10:01.103 "uuid": "b4d4cf77-2359-41b0-8b46-963dd05d8b6e", 00:10:01.103 "is_configured": true, 00:10:01.103 "data_offset": 0, 00:10:01.103 "data_size": 65536 00:10:01.103 }, 00:10:01.103 { 00:10:01.103 "name": "BaseBdev4", 00:10:01.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.103 "is_configured": false, 00:10:01.103 "data_offset": 0, 00:10:01.103 "data_size": 0 00:10:01.103 } 00:10:01.103 ] 00:10:01.103 }' 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.103 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.672 [2024-10-29 10:59:06.881915] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:01.672 [2024-10-29 10:59:06.881978] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:01.672 [2024-10-29 10:59:06.881989] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:01.672 [2024-10-29 10:59:06.882341] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:01.672 [2024-10-29 10:59:06.882528] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:01.672 [2024-10-29 10:59:06.882544] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:01.672 [2024-10-29 10:59:06.882791] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:01.672 BaseBdev4 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.672 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.672 [ 00:10:01.672 { 00:10:01.672 "name": "BaseBdev4", 00:10:01.672 "aliases": [ 00:10:01.672 "0fa46363-1b0f-4c34-b564-d44dab2e0c18" 00:10:01.672 ], 00:10:01.672 "product_name": "Malloc disk", 00:10:01.672 "block_size": 512, 00:10:01.672 "num_blocks": 65536, 00:10:01.672 "uuid": "0fa46363-1b0f-4c34-b564-d44dab2e0c18", 00:10:01.672 "assigned_rate_limits": { 00:10:01.672 "rw_ios_per_sec": 0, 00:10:01.672 "rw_mbytes_per_sec": 0, 00:10:01.672 "r_mbytes_per_sec": 0, 00:10:01.672 "w_mbytes_per_sec": 0 00:10:01.672 }, 00:10:01.672 "claimed": true, 00:10:01.672 "claim_type": "exclusive_write", 00:10:01.672 "zoned": false, 00:10:01.672 "supported_io_types": { 00:10:01.672 "read": true, 00:10:01.672 "write": true, 00:10:01.672 "unmap": true, 00:10:01.672 "flush": true, 00:10:01.672 "reset": true, 00:10:01.672 "nvme_admin": false, 00:10:01.672 "nvme_io": false, 00:10:01.672 "nvme_io_md": false, 00:10:01.672 "write_zeroes": true, 00:10:01.672 "zcopy": true, 00:10:01.672 "get_zone_info": false, 00:10:01.672 "zone_management": false, 00:10:01.672 "zone_append": false, 00:10:01.672 "compare": false, 00:10:01.672 "compare_and_write": false, 00:10:01.672 "abort": true, 00:10:01.672 "seek_hole": false, 00:10:01.672 "seek_data": false, 00:10:01.672 "copy": true, 00:10:01.672 "nvme_iov_md": false 00:10:01.672 }, 00:10:01.672 "memory_domains": [ 00:10:01.672 { 00:10:01.672 "dma_device_id": "system", 00:10:01.672 "dma_device_type": 1 00:10:01.672 }, 00:10:01.672 { 00:10:01.672 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.673 "dma_device_type": 2 00:10:01.673 } 00:10:01.673 ], 00:10:01.673 "driver_specific": {} 00:10:01.673 } 00:10:01.673 ] 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.673 "name": "Existed_Raid", 00:10:01.673 "uuid": "9f338563-cdf3-451e-ad27-e217f5f3a0d3", 00:10:01.673 "strip_size_kb": 64, 00:10:01.673 "state": "online", 00:10:01.673 "raid_level": "raid0", 00:10:01.673 "superblock": false, 00:10:01.673 "num_base_bdevs": 4, 00:10:01.673 "num_base_bdevs_discovered": 4, 00:10:01.673 "num_base_bdevs_operational": 4, 00:10:01.673 "base_bdevs_list": [ 00:10:01.673 { 00:10:01.673 "name": "BaseBdev1", 00:10:01.673 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:10:01.673 "is_configured": true, 00:10:01.673 "data_offset": 0, 00:10:01.673 "data_size": 65536 00:10:01.673 }, 00:10:01.673 { 00:10:01.673 "name": "BaseBdev2", 00:10:01.673 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:01.673 "is_configured": true, 00:10:01.673 "data_offset": 0, 00:10:01.673 "data_size": 65536 00:10:01.673 }, 00:10:01.673 { 00:10:01.673 "name": "BaseBdev3", 00:10:01.673 "uuid": "b4d4cf77-2359-41b0-8b46-963dd05d8b6e", 00:10:01.673 "is_configured": true, 00:10:01.673 "data_offset": 0, 00:10:01.673 "data_size": 65536 00:10:01.673 }, 00:10:01.673 { 00:10:01.673 "name": "BaseBdev4", 00:10:01.673 "uuid": "0fa46363-1b0f-4c34-b564-d44dab2e0c18", 00:10:01.673 "is_configured": true, 00:10:01.673 "data_offset": 0, 00:10:01.673 "data_size": 65536 00:10:01.673 } 00:10:01.673 ] 00:10:01.673 }' 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.673 10:59:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:01.932 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.933 [2024-10-29 10:59:07.397628] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:01.933 "name": "Existed_Raid", 00:10:01.933 "aliases": [ 00:10:01.933 "9f338563-cdf3-451e-ad27-e217f5f3a0d3" 00:10:01.933 ], 00:10:01.933 "product_name": "Raid Volume", 00:10:01.933 "block_size": 512, 00:10:01.933 "num_blocks": 262144, 00:10:01.933 "uuid": "9f338563-cdf3-451e-ad27-e217f5f3a0d3", 00:10:01.933 "assigned_rate_limits": { 00:10:01.933 "rw_ios_per_sec": 0, 00:10:01.933 "rw_mbytes_per_sec": 0, 00:10:01.933 "r_mbytes_per_sec": 0, 00:10:01.933 "w_mbytes_per_sec": 0 00:10:01.933 }, 00:10:01.933 "claimed": false, 00:10:01.933 "zoned": false, 00:10:01.933 "supported_io_types": { 00:10:01.933 "read": true, 00:10:01.933 "write": true, 00:10:01.933 "unmap": true, 00:10:01.933 "flush": true, 00:10:01.933 "reset": true, 00:10:01.933 "nvme_admin": false, 00:10:01.933 "nvme_io": false, 00:10:01.933 "nvme_io_md": false, 00:10:01.933 "write_zeroes": true, 00:10:01.933 "zcopy": false, 00:10:01.933 "get_zone_info": false, 00:10:01.933 "zone_management": false, 00:10:01.933 "zone_append": false, 00:10:01.933 "compare": false, 00:10:01.933 "compare_and_write": false, 00:10:01.933 "abort": false, 00:10:01.933 "seek_hole": false, 00:10:01.933 "seek_data": false, 00:10:01.933 "copy": false, 00:10:01.933 "nvme_iov_md": false 00:10:01.933 }, 00:10:01.933 "memory_domains": [ 00:10:01.933 { 00:10:01.933 "dma_device_id": "system", 00:10:01.933 "dma_device_type": 1 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.933 "dma_device_type": 2 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "system", 00:10:01.933 "dma_device_type": 1 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.933 "dma_device_type": 2 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "system", 00:10:01.933 "dma_device_type": 1 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.933 "dma_device_type": 2 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "system", 00:10:01.933 "dma_device_type": 1 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.933 "dma_device_type": 2 00:10:01.933 } 00:10:01.933 ], 00:10:01.933 "driver_specific": { 00:10:01.933 "raid": { 00:10:01.933 "uuid": "9f338563-cdf3-451e-ad27-e217f5f3a0d3", 00:10:01.933 "strip_size_kb": 64, 00:10:01.933 "state": "online", 00:10:01.933 "raid_level": "raid0", 00:10:01.933 "superblock": false, 00:10:01.933 "num_base_bdevs": 4, 00:10:01.933 "num_base_bdevs_discovered": 4, 00:10:01.933 "num_base_bdevs_operational": 4, 00:10:01.933 "base_bdevs_list": [ 00:10:01.933 { 00:10:01.933 "name": "BaseBdev1", 00:10:01.933 "uuid": "c8edfbd8-6636-4b5d-b18f-9401d481ff53", 00:10:01.933 "is_configured": true, 00:10:01.933 "data_offset": 0, 00:10:01.933 "data_size": 65536 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "name": "BaseBdev2", 00:10:01.933 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:01.933 "is_configured": true, 00:10:01.933 "data_offset": 0, 00:10:01.933 "data_size": 65536 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "name": "BaseBdev3", 00:10:01.933 "uuid": "b4d4cf77-2359-41b0-8b46-963dd05d8b6e", 00:10:01.933 "is_configured": true, 00:10:01.933 "data_offset": 0, 00:10:01.933 "data_size": 65536 00:10:01.933 }, 00:10:01.933 { 00:10:01.933 "name": "BaseBdev4", 00:10:01.933 "uuid": "0fa46363-1b0f-4c34-b564-d44dab2e0c18", 00:10:01.933 "is_configured": true, 00:10:01.933 "data_offset": 0, 00:10:01.933 "data_size": 65536 00:10:01.933 } 00:10:01.933 ] 00:10:01.933 } 00:10:01.933 } 00:10:01.933 }' 00:10:01.933 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:02.194 BaseBdev2 00:10:02.194 BaseBdev3 00:10:02.194 BaseBdev4' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.194 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.454 [2024-10-29 10:59:07.712757] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:02.454 [2024-10-29 10:59:07.712907] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:02.454 [2024-10-29 10:59:07.713004] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.454 "name": "Existed_Raid", 00:10:02.454 "uuid": "9f338563-cdf3-451e-ad27-e217f5f3a0d3", 00:10:02.454 "strip_size_kb": 64, 00:10:02.454 "state": "offline", 00:10:02.454 "raid_level": "raid0", 00:10:02.454 "superblock": false, 00:10:02.454 "num_base_bdevs": 4, 00:10:02.454 "num_base_bdevs_discovered": 3, 00:10:02.454 "num_base_bdevs_operational": 3, 00:10:02.454 "base_bdevs_list": [ 00:10:02.454 { 00:10:02.454 "name": null, 00:10:02.454 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.454 "is_configured": false, 00:10:02.454 "data_offset": 0, 00:10:02.454 "data_size": 65536 00:10:02.454 }, 00:10:02.454 { 00:10:02.454 "name": "BaseBdev2", 00:10:02.454 "uuid": "6a49e1f6-47ff-48b6-99fc-1a1353404588", 00:10:02.454 "is_configured": true, 00:10:02.454 "data_offset": 0, 00:10:02.454 "data_size": 65536 00:10:02.454 }, 00:10:02.454 { 00:10:02.454 "name": "BaseBdev3", 00:10:02.454 "uuid": "b4d4cf77-2359-41b0-8b46-963dd05d8b6e", 00:10:02.454 "is_configured": true, 00:10:02.454 "data_offset": 0, 00:10:02.454 "data_size": 65536 00:10:02.454 }, 00:10:02.454 { 00:10:02.454 "name": "BaseBdev4", 00:10:02.454 "uuid": "0fa46363-1b0f-4c34-b564-d44dab2e0c18", 00:10:02.454 "is_configured": true, 00:10:02.454 "data_offset": 0, 00:10:02.454 "data_size": 65536 00:10:02.454 } 00:10:02.454 ] 00:10:02.454 }' 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.454 10:59:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.714 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.973 [2024-10-29 10:59:08.213557] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:02.973 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 [2024-10-29 10:59:08.295111] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 [2024-10-29 10:59:08.360436] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:02.974 [2024-10-29 10:59:08.360581] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 BaseBdev2 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.974 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.233 [ 00:10:03.233 { 00:10:03.233 "name": "BaseBdev2", 00:10:03.233 "aliases": [ 00:10:03.233 "f9537a5e-576d-477b-8ee1-b1996b690822" 00:10:03.233 ], 00:10:03.233 "product_name": "Malloc disk", 00:10:03.233 "block_size": 512, 00:10:03.233 "num_blocks": 65536, 00:10:03.233 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:03.233 "assigned_rate_limits": { 00:10:03.233 "rw_ios_per_sec": 0, 00:10:03.233 "rw_mbytes_per_sec": 0, 00:10:03.233 "r_mbytes_per_sec": 0, 00:10:03.233 "w_mbytes_per_sec": 0 00:10:03.233 }, 00:10:03.233 "claimed": false, 00:10:03.233 "zoned": false, 00:10:03.233 "supported_io_types": { 00:10:03.233 "read": true, 00:10:03.233 "write": true, 00:10:03.233 "unmap": true, 00:10:03.233 "flush": true, 00:10:03.233 "reset": true, 00:10:03.233 "nvme_admin": false, 00:10:03.233 "nvme_io": false, 00:10:03.233 "nvme_io_md": false, 00:10:03.233 "write_zeroes": true, 00:10:03.233 "zcopy": true, 00:10:03.233 "get_zone_info": false, 00:10:03.233 "zone_management": false, 00:10:03.233 "zone_append": false, 00:10:03.233 "compare": false, 00:10:03.233 "compare_and_write": false, 00:10:03.233 "abort": true, 00:10:03.233 "seek_hole": false, 00:10:03.233 "seek_data": false, 00:10:03.233 "copy": true, 00:10:03.233 "nvme_iov_md": false 00:10:03.233 }, 00:10:03.233 "memory_domains": [ 00:10:03.233 { 00:10:03.233 "dma_device_id": "system", 00:10:03.233 "dma_device_type": 1 00:10:03.233 }, 00:10:03.233 { 00:10:03.233 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.233 "dma_device_type": 2 00:10:03.233 } 00:10:03.233 ], 00:10:03.233 "driver_specific": {} 00:10:03.233 } 00:10:03.233 ] 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.233 BaseBdev3 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.233 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.233 [ 00:10:03.233 { 00:10:03.233 "name": "BaseBdev3", 00:10:03.233 "aliases": [ 00:10:03.233 "97a5befd-fa3d-45df-8798-702a057b00b6" 00:10:03.233 ], 00:10:03.234 "product_name": "Malloc disk", 00:10:03.234 "block_size": 512, 00:10:03.234 "num_blocks": 65536, 00:10:03.234 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:03.234 "assigned_rate_limits": { 00:10:03.234 "rw_ios_per_sec": 0, 00:10:03.234 "rw_mbytes_per_sec": 0, 00:10:03.234 "r_mbytes_per_sec": 0, 00:10:03.234 "w_mbytes_per_sec": 0 00:10:03.234 }, 00:10:03.234 "claimed": false, 00:10:03.234 "zoned": false, 00:10:03.234 "supported_io_types": { 00:10:03.234 "read": true, 00:10:03.234 "write": true, 00:10:03.234 "unmap": true, 00:10:03.234 "flush": true, 00:10:03.234 "reset": true, 00:10:03.234 "nvme_admin": false, 00:10:03.234 "nvme_io": false, 00:10:03.234 "nvme_io_md": false, 00:10:03.234 "write_zeroes": true, 00:10:03.234 "zcopy": true, 00:10:03.234 "get_zone_info": false, 00:10:03.234 "zone_management": false, 00:10:03.234 "zone_append": false, 00:10:03.234 "compare": false, 00:10:03.234 "compare_and_write": false, 00:10:03.234 "abort": true, 00:10:03.234 "seek_hole": false, 00:10:03.234 "seek_data": false, 00:10:03.234 "copy": true, 00:10:03.234 "nvme_iov_md": false 00:10:03.234 }, 00:10:03.234 "memory_domains": [ 00:10:03.234 { 00:10:03.234 "dma_device_id": "system", 00:10:03.234 "dma_device_type": 1 00:10:03.234 }, 00:10:03.234 { 00:10:03.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.234 "dma_device_type": 2 00:10:03.234 } 00:10:03.234 ], 00:10:03.234 "driver_specific": {} 00:10:03.234 } 00:10:03.234 ] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.234 BaseBdev4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.234 [ 00:10:03.234 { 00:10:03.234 "name": "BaseBdev4", 00:10:03.234 "aliases": [ 00:10:03.234 "b4518916-38d3-4e1e-adaa-20815fc6dfb3" 00:10:03.234 ], 00:10:03.234 "product_name": "Malloc disk", 00:10:03.234 "block_size": 512, 00:10:03.234 "num_blocks": 65536, 00:10:03.234 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:03.234 "assigned_rate_limits": { 00:10:03.234 "rw_ios_per_sec": 0, 00:10:03.234 "rw_mbytes_per_sec": 0, 00:10:03.234 "r_mbytes_per_sec": 0, 00:10:03.234 "w_mbytes_per_sec": 0 00:10:03.234 }, 00:10:03.234 "claimed": false, 00:10:03.234 "zoned": false, 00:10:03.234 "supported_io_types": { 00:10:03.234 "read": true, 00:10:03.234 "write": true, 00:10:03.234 "unmap": true, 00:10:03.234 "flush": true, 00:10:03.234 "reset": true, 00:10:03.234 "nvme_admin": false, 00:10:03.234 "nvme_io": false, 00:10:03.234 "nvme_io_md": false, 00:10:03.234 "write_zeroes": true, 00:10:03.234 "zcopy": true, 00:10:03.234 "get_zone_info": false, 00:10:03.234 "zone_management": false, 00:10:03.234 "zone_append": false, 00:10:03.234 "compare": false, 00:10:03.234 "compare_and_write": false, 00:10:03.234 "abort": true, 00:10:03.234 "seek_hole": false, 00:10:03.234 "seek_data": false, 00:10:03.234 "copy": true, 00:10:03.234 "nvme_iov_md": false 00:10:03.234 }, 00:10:03.234 "memory_domains": [ 00:10:03.234 { 00:10:03.234 "dma_device_id": "system", 00:10:03.234 "dma_device_type": 1 00:10:03.234 }, 00:10:03.234 { 00:10:03.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.234 "dma_device_type": 2 00:10:03.234 } 00:10:03.234 ], 00:10:03.234 "driver_specific": {} 00:10:03.234 } 00:10:03.234 ] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.234 [2024-10-29 10:59:08.623715] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:03.234 [2024-10-29 10:59:08.623853] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:03.234 [2024-10-29 10:59:08.623912] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:03.234 [2024-10-29 10:59:08.626343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:03.234 [2024-10-29 10:59:08.626469] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.234 "name": "Existed_Raid", 00:10:03.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.234 "strip_size_kb": 64, 00:10:03.234 "state": "configuring", 00:10:03.234 "raid_level": "raid0", 00:10:03.234 "superblock": false, 00:10:03.234 "num_base_bdevs": 4, 00:10:03.234 "num_base_bdevs_discovered": 3, 00:10:03.234 "num_base_bdevs_operational": 4, 00:10:03.234 "base_bdevs_list": [ 00:10:03.234 { 00:10:03.234 "name": "BaseBdev1", 00:10:03.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.234 "is_configured": false, 00:10:03.234 "data_offset": 0, 00:10:03.234 "data_size": 0 00:10:03.234 }, 00:10:03.234 { 00:10:03.234 "name": "BaseBdev2", 00:10:03.234 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:03.234 "is_configured": true, 00:10:03.234 "data_offset": 0, 00:10:03.234 "data_size": 65536 00:10:03.234 }, 00:10:03.234 { 00:10:03.234 "name": "BaseBdev3", 00:10:03.234 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:03.234 "is_configured": true, 00:10:03.234 "data_offset": 0, 00:10:03.234 "data_size": 65536 00:10:03.234 }, 00:10:03.234 { 00:10:03.234 "name": "BaseBdev4", 00:10:03.234 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:03.234 "is_configured": true, 00:10:03.234 "data_offset": 0, 00:10:03.234 "data_size": 65536 00:10:03.234 } 00:10:03.234 ] 00:10:03.234 }' 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.234 10:59:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.800 [2024-10-29 10:59:09.059460] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.800 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.801 "name": "Existed_Raid", 00:10:03.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.801 "strip_size_kb": 64, 00:10:03.801 "state": "configuring", 00:10:03.801 "raid_level": "raid0", 00:10:03.801 "superblock": false, 00:10:03.801 "num_base_bdevs": 4, 00:10:03.801 "num_base_bdevs_discovered": 2, 00:10:03.801 "num_base_bdevs_operational": 4, 00:10:03.801 "base_bdevs_list": [ 00:10:03.801 { 00:10:03.801 "name": "BaseBdev1", 00:10:03.801 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.801 "is_configured": false, 00:10:03.801 "data_offset": 0, 00:10:03.801 "data_size": 0 00:10:03.801 }, 00:10:03.801 { 00:10:03.801 "name": null, 00:10:03.801 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:03.801 "is_configured": false, 00:10:03.801 "data_offset": 0, 00:10:03.801 "data_size": 65536 00:10:03.801 }, 00:10:03.801 { 00:10:03.801 "name": "BaseBdev3", 00:10:03.801 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:03.801 "is_configured": true, 00:10:03.801 "data_offset": 0, 00:10:03.801 "data_size": 65536 00:10:03.801 }, 00:10:03.801 { 00:10:03.801 "name": "BaseBdev4", 00:10:03.801 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:03.801 "is_configured": true, 00:10:03.801 "data_offset": 0, 00:10:03.801 "data_size": 65536 00:10:03.801 } 00:10:03.801 ] 00:10:03.801 }' 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.801 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.060 [2024-10-29 10:59:09.551933] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:04.060 BaseBdev1 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.060 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.320 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.320 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:04.320 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.320 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.320 [ 00:10:04.320 { 00:10:04.320 "name": "BaseBdev1", 00:10:04.320 "aliases": [ 00:10:04.320 "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c" 00:10:04.320 ], 00:10:04.320 "product_name": "Malloc disk", 00:10:04.320 "block_size": 512, 00:10:04.320 "num_blocks": 65536, 00:10:04.320 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:04.320 "assigned_rate_limits": { 00:10:04.320 "rw_ios_per_sec": 0, 00:10:04.320 "rw_mbytes_per_sec": 0, 00:10:04.320 "r_mbytes_per_sec": 0, 00:10:04.320 "w_mbytes_per_sec": 0 00:10:04.320 }, 00:10:04.320 "claimed": true, 00:10:04.320 "claim_type": "exclusive_write", 00:10:04.320 "zoned": false, 00:10:04.320 "supported_io_types": { 00:10:04.320 "read": true, 00:10:04.320 "write": true, 00:10:04.320 "unmap": true, 00:10:04.320 "flush": true, 00:10:04.320 "reset": true, 00:10:04.320 "nvme_admin": false, 00:10:04.320 "nvme_io": false, 00:10:04.320 "nvme_io_md": false, 00:10:04.320 "write_zeroes": true, 00:10:04.320 "zcopy": true, 00:10:04.320 "get_zone_info": false, 00:10:04.320 "zone_management": false, 00:10:04.320 "zone_append": false, 00:10:04.320 "compare": false, 00:10:04.320 "compare_and_write": false, 00:10:04.320 "abort": true, 00:10:04.320 "seek_hole": false, 00:10:04.320 "seek_data": false, 00:10:04.320 "copy": true, 00:10:04.320 "nvme_iov_md": false 00:10:04.320 }, 00:10:04.320 "memory_domains": [ 00:10:04.320 { 00:10:04.320 "dma_device_id": "system", 00:10:04.320 "dma_device_type": 1 00:10:04.320 }, 00:10:04.320 { 00:10:04.320 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.320 "dma_device_type": 2 00:10:04.320 } 00:10:04.320 ], 00:10:04.320 "driver_specific": {} 00:10:04.320 } 00:10:04.320 ] 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.321 "name": "Existed_Raid", 00:10:04.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.321 "strip_size_kb": 64, 00:10:04.321 "state": "configuring", 00:10:04.321 "raid_level": "raid0", 00:10:04.321 "superblock": false, 00:10:04.321 "num_base_bdevs": 4, 00:10:04.321 "num_base_bdevs_discovered": 3, 00:10:04.321 "num_base_bdevs_operational": 4, 00:10:04.321 "base_bdevs_list": [ 00:10:04.321 { 00:10:04.321 "name": "BaseBdev1", 00:10:04.321 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:04.321 "is_configured": true, 00:10:04.321 "data_offset": 0, 00:10:04.321 "data_size": 65536 00:10:04.321 }, 00:10:04.321 { 00:10:04.321 "name": null, 00:10:04.321 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:04.321 "is_configured": false, 00:10:04.321 "data_offset": 0, 00:10:04.321 "data_size": 65536 00:10:04.321 }, 00:10:04.321 { 00:10:04.321 "name": "BaseBdev3", 00:10:04.321 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:04.321 "is_configured": true, 00:10:04.321 "data_offset": 0, 00:10:04.321 "data_size": 65536 00:10:04.321 }, 00:10:04.321 { 00:10:04.321 "name": "BaseBdev4", 00:10:04.321 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:04.321 "is_configured": true, 00:10:04.321 "data_offset": 0, 00:10:04.321 "data_size": 65536 00:10:04.321 } 00:10:04.321 ] 00:10:04.321 }' 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.321 10:59:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.578 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:04.578 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.578 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.578 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.837 [2024-10-29 10:59:10.115571] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.837 "name": "Existed_Raid", 00:10:04.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.837 "strip_size_kb": 64, 00:10:04.837 "state": "configuring", 00:10:04.837 "raid_level": "raid0", 00:10:04.837 "superblock": false, 00:10:04.837 "num_base_bdevs": 4, 00:10:04.837 "num_base_bdevs_discovered": 2, 00:10:04.837 "num_base_bdevs_operational": 4, 00:10:04.837 "base_bdevs_list": [ 00:10:04.837 { 00:10:04.837 "name": "BaseBdev1", 00:10:04.837 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:04.837 "is_configured": true, 00:10:04.837 "data_offset": 0, 00:10:04.837 "data_size": 65536 00:10:04.837 }, 00:10:04.837 { 00:10:04.837 "name": null, 00:10:04.837 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:04.837 "is_configured": false, 00:10:04.837 "data_offset": 0, 00:10:04.837 "data_size": 65536 00:10:04.837 }, 00:10:04.837 { 00:10:04.837 "name": null, 00:10:04.837 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:04.837 "is_configured": false, 00:10:04.837 "data_offset": 0, 00:10:04.837 "data_size": 65536 00:10:04.837 }, 00:10:04.837 { 00:10:04.837 "name": "BaseBdev4", 00:10:04.837 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:04.837 "is_configured": true, 00:10:04.837 "data_offset": 0, 00:10:04.837 "data_size": 65536 00:10:04.837 } 00:10:04.837 ] 00:10:04.837 }' 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.837 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.095 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.353 [2024-10-29 10:59:10.595600] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.353 "name": "Existed_Raid", 00:10:05.353 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:05.353 "strip_size_kb": 64, 00:10:05.353 "state": "configuring", 00:10:05.353 "raid_level": "raid0", 00:10:05.353 "superblock": false, 00:10:05.353 "num_base_bdevs": 4, 00:10:05.353 "num_base_bdevs_discovered": 3, 00:10:05.353 "num_base_bdevs_operational": 4, 00:10:05.353 "base_bdevs_list": [ 00:10:05.353 { 00:10:05.353 "name": "BaseBdev1", 00:10:05.353 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:05.353 "is_configured": true, 00:10:05.353 "data_offset": 0, 00:10:05.353 "data_size": 65536 00:10:05.353 }, 00:10:05.353 { 00:10:05.353 "name": null, 00:10:05.353 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:05.353 "is_configured": false, 00:10:05.353 "data_offset": 0, 00:10:05.353 "data_size": 65536 00:10:05.353 }, 00:10:05.353 { 00:10:05.353 "name": "BaseBdev3", 00:10:05.353 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:05.353 "is_configured": true, 00:10:05.353 "data_offset": 0, 00:10:05.353 "data_size": 65536 00:10:05.353 }, 00:10:05.353 { 00:10:05.353 "name": "BaseBdev4", 00:10:05.353 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:05.353 "is_configured": true, 00:10:05.353 "data_offset": 0, 00:10:05.353 "data_size": 65536 00:10:05.353 } 00:10:05.353 ] 00:10:05.353 }' 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.353 10:59:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.612 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.612 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.612 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.612 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:05.612 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.869 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:05.869 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.870 [2024-10-29 10:59:11.123619] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.870 "name": "Existed_Raid", 00:10:05.870 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:05.870 "strip_size_kb": 64, 00:10:05.870 "state": "configuring", 00:10:05.870 "raid_level": "raid0", 00:10:05.870 "superblock": false, 00:10:05.870 "num_base_bdevs": 4, 00:10:05.870 "num_base_bdevs_discovered": 2, 00:10:05.870 "num_base_bdevs_operational": 4, 00:10:05.870 "base_bdevs_list": [ 00:10:05.870 { 00:10:05.870 "name": null, 00:10:05.870 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:05.870 "is_configured": false, 00:10:05.870 "data_offset": 0, 00:10:05.870 "data_size": 65536 00:10:05.870 }, 00:10:05.870 { 00:10:05.870 "name": null, 00:10:05.870 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:05.870 "is_configured": false, 00:10:05.870 "data_offset": 0, 00:10:05.870 "data_size": 65536 00:10:05.870 }, 00:10:05.870 { 00:10:05.870 "name": "BaseBdev3", 00:10:05.870 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:05.870 "is_configured": true, 00:10:05.870 "data_offset": 0, 00:10:05.870 "data_size": 65536 00:10:05.870 }, 00:10:05.870 { 00:10:05.870 "name": "BaseBdev4", 00:10:05.870 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:05.870 "is_configured": true, 00:10:05.870 "data_offset": 0, 00:10:05.870 "data_size": 65536 00:10:05.870 } 00:10:05.870 ] 00:10:05.870 }' 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.870 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.128 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.128 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:06.128 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.128 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.128 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.387 [2024-10-29 10:59:11.635623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.387 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.388 "name": "Existed_Raid", 00:10:06.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.388 "strip_size_kb": 64, 00:10:06.388 "state": "configuring", 00:10:06.388 "raid_level": "raid0", 00:10:06.388 "superblock": false, 00:10:06.388 "num_base_bdevs": 4, 00:10:06.388 "num_base_bdevs_discovered": 3, 00:10:06.388 "num_base_bdevs_operational": 4, 00:10:06.388 "base_bdevs_list": [ 00:10:06.388 { 00:10:06.388 "name": null, 00:10:06.388 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:06.388 "is_configured": false, 00:10:06.388 "data_offset": 0, 00:10:06.388 "data_size": 65536 00:10:06.388 }, 00:10:06.388 { 00:10:06.388 "name": "BaseBdev2", 00:10:06.388 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:06.388 "is_configured": true, 00:10:06.388 "data_offset": 0, 00:10:06.388 "data_size": 65536 00:10:06.388 }, 00:10:06.388 { 00:10:06.388 "name": "BaseBdev3", 00:10:06.388 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:06.388 "is_configured": true, 00:10:06.388 "data_offset": 0, 00:10:06.388 "data_size": 65536 00:10:06.388 }, 00:10:06.388 { 00:10:06.388 "name": "BaseBdev4", 00:10:06.388 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:06.388 "is_configured": true, 00:10:06.388 "data_offset": 0, 00:10:06.388 "data_size": 65536 00:10:06.388 } 00:10:06.388 ] 00:10:06.388 }' 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.388 10:59:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.647 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.907 [2024-10-29 10:59:12.193254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:06.907 [2024-10-29 10:59:12.193393] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:06.907 [2024-10-29 10:59:12.193449] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:06.907 [2024-10-29 10:59:12.193881] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:06.907 [2024-10-29 10:59:12.194092] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:06.907 [2024-10-29 10:59:12.194143] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:06.907 [2024-10-29 10:59:12.194480] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:06.907 NewBaseBdev 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.907 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.907 [ 00:10:06.907 { 00:10:06.907 "name": "NewBaseBdev", 00:10:06.907 "aliases": [ 00:10:06.907 "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c" 00:10:06.907 ], 00:10:06.907 "product_name": "Malloc disk", 00:10:06.908 "block_size": 512, 00:10:06.908 "num_blocks": 65536, 00:10:06.908 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:06.908 "assigned_rate_limits": { 00:10:06.908 "rw_ios_per_sec": 0, 00:10:06.908 "rw_mbytes_per_sec": 0, 00:10:06.908 "r_mbytes_per_sec": 0, 00:10:06.908 "w_mbytes_per_sec": 0 00:10:06.908 }, 00:10:06.908 "claimed": true, 00:10:06.908 "claim_type": "exclusive_write", 00:10:06.908 "zoned": false, 00:10:06.908 "supported_io_types": { 00:10:06.908 "read": true, 00:10:06.908 "write": true, 00:10:06.908 "unmap": true, 00:10:06.908 "flush": true, 00:10:06.908 "reset": true, 00:10:06.908 "nvme_admin": false, 00:10:06.908 "nvme_io": false, 00:10:06.908 "nvme_io_md": false, 00:10:06.908 "write_zeroes": true, 00:10:06.908 "zcopy": true, 00:10:06.908 "get_zone_info": false, 00:10:06.908 "zone_management": false, 00:10:06.908 "zone_append": false, 00:10:06.908 "compare": false, 00:10:06.908 "compare_and_write": false, 00:10:06.908 "abort": true, 00:10:06.908 "seek_hole": false, 00:10:06.908 "seek_data": false, 00:10:06.908 "copy": true, 00:10:06.908 "nvme_iov_md": false 00:10:06.908 }, 00:10:06.908 "memory_domains": [ 00:10:06.908 { 00:10:06.908 "dma_device_id": "system", 00:10:06.908 "dma_device_type": 1 00:10:06.908 }, 00:10:06.908 { 00:10:06.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.908 "dma_device_type": 2 00:10:06.908 } 00:10:06.908 ], 00:10:06.908 "driver_specific": {} 00:10:06.908 } 00:10:06.908 ] 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.908 "name": "Existed_Raid", 00:10:06.908 "uuid": "d018c2fc-3640-4fa6-8e41-a6f096169fe1", 00:10:06.908 "strip_size_kb": 64, 00:10:06.908 "state": "online", 00:10:06.908 "raid_level": "raid0", 00:10:06.908 "superblock": false, 00:10:06.908 "num_base_bdevs": 4, 00:10:06.908 "num_base_bdevs_discovered": 4, 00:10:06.908 "num_base_bdevs_operational": 4, 00:10:06.908 "base_bdevs_list": [ 00:10:06.908 { 00:10:06.908 "name": "NewBaseBdev", 00:10:06.908 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:06.908 "is_configured": true, 00:10:06.908 "data_offset": 0, 00:10:06.908 "data_size": 65536 00:10:06.908 }, 00:10:06.908 { 00:10:06.908 "name": "BaseBdev2", 00:10:06.908 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:06.908 "is_configured": true, 00:10:06.908 "data_offset": 0, 00:10:06.908 "data_size": 65536 00:10:06.908 }, 00:10:06.908 { 00:10:06.908 "name": "BaseBdev3", 00:10:06.908 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:06.908 "is_configured": true, 00:10:06.908 "data_offset": 0, 00:10:06.908 "data_size": 65536 00:10:06.908 }, 00:10:06.908 { 00:10:06.908 "name": "BaseBdev4", 00:10:06.908 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:06.908 "is_configured": true, 00:10:06.908 "data_offset": 0, 00:10:06.908 "data_size": 65536 00:10:06.908 } 00:10:06.908 ] 00:10:06.908 }' 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.908 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.477 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.478 [2024-10-29 10:59:12.692998] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:07.478 "name": "Existed_Raid", 00:10:07.478 "aliases": [ 00:10:07.478 "d018c2fc-3640-4fa6-8e41-a6f096169fe1" 00:10:07.478 ], 00:10:07.478 "product_name": "Raid Volume", 00:10:07.478 "block_size": 512, 00:10:07.478 "num_blocks": 262144, 00:10:07.478 "uuid": "d018c2fc-3640-4fa6-8e41-a6f096169fe1", 00:10:07.478 "assigned_rate_limits": { 00:10:07.478 "rw_ios_per_sec": 0, 00:10:07.478 "rw_mbytes_per_sec": 0, 00:10:07.478 "r_mbytes_per_sec": 0, 00:10:07.478 "w_mbytes_per_sec": 0 00:10:07.478 }, 00:10:07.478 "claimed": false, 00:10:07.478 "zoned": false, 00:10:07.478 "supported_io_types": { 00:10:07.478 "read": true, 00:10:07.478 "write": true, 00:10:07.478 "unmap": true, 00:10:07.478 "flush": true, 00:10:07.478 "reset": true, 00:10:07.478 "nvme_admin": false, 00:10:07.478 "nvme_io": false, 00:10:07.478 "nvme_io_md": false, 00:10:07.478 "write_zeroes": true, 00:10:07.478 "zcopy": false, 00:10:07.478 "get_zone_info": false, 00:10:07.478 "zone_management": false, 00:10:07.478 "zone_append": false, 00:10:07.478 "compare": false, 00:10:07.478 "compare_and_write": false, 00:10:07.478 "abort": false, 00:10:07.478 "seek_hole": false, 00:10:07.478 "seek_data": false, 00:10:07.478 "copy": false, 00:10:07.478 "nvme_iov_md": false 00:10:07.478 }, 00:10:07.478 "memory_domains": [ 00:10:07.478 { 00:10:07.478 "dma_device_id": "system", 00:10:07.478 "dma_device_type": 1 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.478 "dma_device_type": 2 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "system", 00:10:07.478 "dma_device_type": 1 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.478 "dma_device_type": 2 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "system", 00:10:07.478 "dma_device_type": 1 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.478 "dma_device_type": 2 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "system", 00:10:07.478 "dma_device_type": 1 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.478 "dma_device_type": 2 00:10:07.478 } 00:10:07.478 ], 00:10:07.478 "driver_specific": { 00:10:07.478 "raid": { 00:10:07.478 "uuid": "d018c2fc-3640-4fa6-8e41-a6f096169fe1", 00:10:07.478 "strip_size_kb": 64, 00:10:07.478 "state": "online", 00:10:07.478 "raid_level": "raid0", 00:10:07.478 "superblock": false, 00:10:07.478 "num_base_bdevs": 4, 00:10:07.478 "num_base_bdevs_discovered": 4, 00:10:07.478 "num_base_bdevs_operational": 4, 00:10:07.478 "base_bdevs_list": [ 00:10:07.478 { 00:10:07.478 "name": "NewBaseBdev", 00:10:07.478 "uuid": "0ef0e5e8-e1c4-46c4-b68c-38c0e22cf75c", 00:10:07.478 "is_configured": true, 00:10:07.478 "data_offset": 0, 00:10:07.478 "data_size": 65536 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "name": "BaseBdev2", 00:10:07.478 "uuid": "f9537a5e-576d-477b-8ee1-b1996b690822", 00:10:07.478 "is_configured": true, 00:10:07.478 "data_offset": 0, 00:10:07.478 "data_size": 65536 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "name": "BaseBdev3", 00:10:07.478 "uuid": "97a5befd-fa3d-45df-8798-702a057b00b6", 00:10:07.478 "is_configured": true, 00:10:07.478 "data_offset": 0, 00:10:07.478 "data_size": 65536 00:10:07.478 }, 00:10:07.478 { 00:10:07.478 "name": "BaseBdev4", 00:10:07.478 "uuid": "b4518916-38d3-4e1e-adaa-20815fc6dfb3", 00:10:07.478 "is_configured": true, 00:10:07.478 "data_offset": 0, 00:10:07.478 "data_size": 65536 00:10:07.478 } 00:10:07.478 ] 00:10:07.478 } 00:10:07.478 } 00:10:07.478 }' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:07.478 BaseBdev2 00:10:07.478 BaseBdev3 00:10:07.478 BaseBdev4' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.478 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.737 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:07.737 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:07.737 10:59:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:07.737 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.737 10:59:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.737 [2024-10-29 10:59:13.004063] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:07.737 [2024-10-29 10:59:13.004197] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:07.737 [2024-10-29 10:59:13.004369] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:07.737 [2024-10-29 10:59:13.004479] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:07.737 [2024-10-29 10:59:13.004494] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:07.737 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.737 10:59:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80499 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 80499 ']' 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 80499 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 80499 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 80499' 00:10:07.738 killing process with pid 80499 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 80499 00:10:07.738 [2024-10-29 10:59:13.050331] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:07.738 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 80499 00:10:07.738 [2024-10-29 10:59:13.129775] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:07.997 ************************************ 00:10:07.997 END TEST raid_state_function_test 00:10:07.997 ************************************ 00:10:07.997 10:59:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:07.997 00:10:07.997 real 0m9.968s 00:10:07.997 user 0m16.738s 00:10:07.997 sys 0m2.119s 00:10:07.997 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:07.997 10:59:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.257 10:59:13 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:10:08.257 10:59:13 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:08.257 10:59:13 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:08.257 10:59:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:08.257 ************************************ 00:10:08.257 START TEST raid_state_function_test_sb 00:10:08.257 ************************************ 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid0 4 true 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81159 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81159' 00:10:08.257 Process raid pid: 81159 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81159 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 81159 ']' 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:08.257 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:08.257 10:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.257 [2024-10-29 10:59:13.622876] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:08.257 [2024-10-29 10:59:13.623075] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:08.518 [2024-10-29 10:59:13.795190] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:08.518 [2024-10-29 10:59:13.838590] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:08.518 [2024-10-29 10:59:13.918475] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:08.518 [2024-10-29 10:59:13.918630] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:09.087 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:09.087 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:10:09.087 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.088 [2024-10-29 10:59:14.491765] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:09.088 [2024-10-29 10:59:14.491940] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:09.088 [2024-10-29 10:59:14.491965] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:09.088 [2024-10-29 10:59:14.491979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:09.088 [2024-10-29 10:59:14.491986] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:09.088 [2024-10-29 10:59:14.492001] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:09.088 [2024-10-29 10:59:14.492008] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:09.088 [2024-10-29 10:59:14.492018] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.088 "name": "Existed_Raid", 00:10:09.088 "uuid": "e0f0ddcf-8996-4eff-b198-c16516370817", 00:10:09.088 "strip_size_kb": 64, 00:10:09.088 "state": "configuring", 00:10:09.088 "raid_level": "raid0", 00:10:09.088 "superblock": true, 00:10:09.088 "num_base_bdevs": 4, 00:10:09.088 "num_base_bdevs_discovered": 0, 00:10:09.088 "num_base_bdevs_operational": 4, 00:10:09.088 "base_bdevs_list": [ 00:10:09.088 { 00:10:09.088 "name": "BaseBdev1", 00:10:09.088 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.088 "is_configured": false, 00:10:09.088 "data_offset": 0, 00:10:09.088 "data_size": 0 00:10:09.088 }, 00:10:09.088 { 00:10:09.088 "name": "BaseBdev2", 00:10:09.088 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.088 "is_configured": false, 00:10:09.088 "data_offset": 0, 00:10:09.088 "data_size": 0 00:10:09.088 }, 00:10:09.088 { 00:10:09.088 "name": "BaseBdev3", 00:10:09.088 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.088 "is_configured": false, 00:10:09.088 "data_offset": 0, 00:10:09.088 "data_size": 0 00:10:09.088 }, 00:10:09.088 { 00:10:09.088 "name": "BaseBdev4", 00:10:09.088 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.088 "is_configured": false, 00:10:09.088 "data_offset": 0, 00:10:09.088 "data_size": 0 00:10:09.088 } 00:10:09.088 ] 00:10:09.088 }' 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.088 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 [2024-10-29 10:59:14.951578] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:09.658 [2024-10-29 10:59:14.951738] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 [2024-10-29 10:59:14.963605] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:09.658 [2024-10-29 10:59:14.963731] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:09.658 [2024-10-29 10:59:14.963764] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:09.658 [2024-10-29 10:59:14.963796] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:09.658 [2024-10-29 10:59:14.963819] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:09.658 [2024-10-29 10:59:14.963854] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:09.658 [2024-10-29 10:59:14.963877] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:09.658 [2024-10-29 10:59:14.963952] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 [2024-10-29 10:59:14.991571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:09.658 BaseBdev1 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 [ 00:10:09.658 { 00:10:09.658 "name": "BaseBdev1", 00:10:09.658 "aliases": [ 00:10:09.658 "31cefd6c-deda-444c-986f-6ec8375d8abe" 00:10:09.658 ], 00:10:09.658 "product_name": "Malloc disk", 00:10:09.658 "block_size": 512, 00:10:09.658 "num_blocks": 65536, 00:10:09.658 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:09.658 "assigned_rate_limits": { 00:10:09.658 "rw_ios_per_sec": 0, 00:10:09.658 "rw_mbytes_per_sec": 0, 00:10:09.658 "r_mbytes_per_sec": 0, 00:10:09.658 "w_mbytes_per_sec": 0 00:10:09.658 }, 00:10:09.658 "claimed": true, 00:10:09.658 "claim_type": "exclusive_write", 00:10:09.658 "zoned": false, 00:10:09.658 "supported_io_types": { 00:10:09.658 "read": true, 00:10:09.658 "write": true, 00:10:09.658 "unmap": true, 00:10:09.658 "flush": true, 00:10:09.658 "reset": true, 00:10:09.658 "nvme_admin": false, 00:10:09.658 "nvme_io": false, 00:10:09.658 "nvme_io_md": false, 00:10:09.658 "write_zeroes": true, 00:10:09.658 "zcopy": true, 00:10:09.658 "get_zone_info": false, 00:10:09.658 "zone_management": false, 00:10:09.658 "zone_append": false, 00:10:09.658 "compare": false, 00:10:09.658 "compare_and_write": false, 00:10:09.658 "abort": true, 00:10:09.658 "seek_hole": false, 00:10:09.658 "seek_data": false, 00:10:09.658 "copy": true, 00:10:09.658 "nvme_iov_md": false 00:10:09.658 }, 00:10:09.658 "memory_domains": [ 00:10:09.658 { 00:10:09.658 "dma_device_id": "system", 00:10:09.658 "dma_device_type": 1 00:10:09.658 }, 00:10:09.658 { 00:10:09.658 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.658 "dma_device_type": 2 00:10:09.658 } 00:10:09.658 ], 00:10:09.658 "driver_specific": {} 00:10:09.658 } 00:10:09.658 ] 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.658 "name": "Existed_Raid", 00:10:09.658 "uuid": "4782beaf-65da-49ab-bf83-334118c1df6c", 00:10:09.658 "strip_size_kb": 64, 00:10:09.658 "state": "configuring", 00:10:09.658 "raid_level": "raid0", 00:10:09.658 "superblock": true, 00:10:09.658 "num_base_bdevs": 4, 00:10:09.658 "num_base_bdevs_discovered": 1, 00:10:09.658 "num_base_bdevs_operational": 4, 00:10:09.658 "base_bdevs_list": [ 00:10:09.658 { 00:10:09.658 "name": "BaseBdev1", 00:10:09.658 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:09.658 "is_configured": true, 00:10:09.658 "data_offset": 2048, 00:10:09.658 "data_size": 63488 00:10:09.658 }, 00:10:09.658 { 00:10:09.658 "name": "BaseBdev2", 00:10:09.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.658 "is_configured": false, 00:10:09.658 "data_offset": 0, 00:10:09.658 "data_size": 0 00:10:09.658 }, 00:10:09.658 { 00:10:09.658 "name": "BaseBdev3", 00:10:09.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.658 "is_configured": false, 00:10:09.658 "data_offset": 0, 00:10:09.658 "data_size": 0 00:10:09.658 }, 00:10:09.658 { 00:10:09.658 "name": "BaseBdev4", 00:10:09.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.658 "is_configured": false, 00:10:09.658 "data_offset": 0, 00:10:09.658 "data_size": 0 00:10:09.658 } 00:10:09.658 ] 00:10:09.658 }' 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.658 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.235 [2024-10-29 10:59:15.487208] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:10.235 [2024-10-29 10:59:15.487405] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.235 [2024-10-29 10:59:15.499208] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:10.235 [2024-10-29 10:59:15.501492] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:10.235 [2024-10-29 10:59:15.501535] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:10.235 [2024-10-29 10:59:15.501544] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:10.235 [2024-10-29 10:59:15.501552] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:10.235 [2024-10-29 10:59:15.501559] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:10.235 [2024-10-29 10:59:15.501568] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.235 "name": "Existed_Raid", 00:10:10.235 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:10.235 "strip_size_kb": 64, 00:10:10.235 "state": "configuring", 00:10:10.235 "raid_level": "raid0", 00:10:10.235 "superblock": true, 00:10:10.235 "num_base_bdevs": 4, 00:10:10.235 "num_base_bdevs_discovered": 1, 00:10:10.235 "num_base_bdevs_operational": 4, 00:10:10.235 "base_bdevs_list": [ 00:10:10.235 { 00:10:10.235 "name": "BaseBdev1", 00:10:10.235 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:10.235 "is_configured": true, 00:10:10.235 "data_offset": 2048, 00:10:10.235 "data_size": 63488 00:10:10.235 }, 00:10:10.235 { 00:10:10.235 "name": "BaseBdev2", 00:10:10.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.235 "is_configured": false, 00:10:10.235 "data_offset": 0, 00:10:10.235 "data_size": 0 00:10:10.235 }, 00:10:10.235 { 00:10:10.235 "name": "BaseBdev3", 00:10:10.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.235 "is_configured": false, 00:10:10.235 "data_offset": 0, 00:10:10.235 "data_size": 0 00:10:10.235 }, 00:10:10.235 { 00:10:10.235 "name": "BaseBdev4", 00:10:10.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.235 "is_configured": false, 00:10:10.235 "data_offset": 0, 00:10:10.235 "data_size": 0 00:10:10.235 } 00:10:10.235 ] 00:10:10.235 }' 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.235 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.510 10:59:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:10.510 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.510 10:59:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.775 [2024-10-29 10:59:16.011487] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:10.775 BaseBdev2 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.775 [ 00:10:10.775 { 00:10:10.775 "name": "BaseBdev2", 00:10:10.775 "aliases": [ 00:10:10.775 "14228ef6-78b2-43e5-9c88-0bd788d0ff86" 00:10:10.775 ], 00:10:10.775 "product_name": "Malloc disk", 00:10:10.775 "block_size": 512, 00:10:10.775 "num_blocks": 65536, 00:10:10.775 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:10.775 "assigned_rate_limits": { 00:10:10.775 "rw_ios_per_sec": 0, 00:10:10.775 "rw_mbytes_per_sec": 0, 00:10:10.775 "r_mbytes_per_sec": 0, 00:10:10.775 "w_mbytes_per_sec": 0 00:10:10.775 }, 00:10:10.775 "claimed": true, 00:10:10.775 "claim_type": "exclusive_write", 00:10:10.775 "zoned": false, 00:10:10.775 "supported_io_types": { 00:10:10.775 "read": true, 00:10:10.775 "write": true, 00:10:10.775 "unmap": true, 00:10:10.775 "flush": true, 00:10:10.775 "reset": true, 00:10:10.775 "nvme_admin": false, 00:10:10.775 "nvme_io": false, 00:10:10.775 "nvme_io_md": false, 00:10:10.775 "write_zeroes": true, 00:10:10.775 "zcopy": true, 00:10:10.775 "get_zone_info": false, 00:10:10.775 "zone_management": false, 00:10:10.775 "zone_append": false, 00:10:10.775 "compare": false, 00:10:10.775 "compare_and_write": false, 00:10:10.775 "abort": true, 00:10:10.775 "seek_hole": false, 00:10:10.775 "seek_data": false, 00:10:10.775 "copy": true, 00:10:10.775 "nvme_iov_md": false 00:10:10.775 }, 00:10:10.775 "memory_domains": [ 00:10:10.775 { 00:10:10.775 "dma_device_id": "system", 00:10:10.775 "dma_device_type": 1 00:10:10.775 }, 00:10:10.775 { 00:10:10.775 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.775 "dma_device_type": 2 00:10:10.775 } 00:10:10.775 ], 00:10:10.775 "driver_specific": {} 00:10:10.775 } 00:10:10.775 ] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.775 "name": "Existed_Raid", 00:10:10.775 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:10.775 "strip_size_kb": 64, 00:10:10.775 "state": "configuring", 00:10:10.775 "raid_level": "raid0", 00:10:10.775 "superblock": true, 00:10:10.775 "num_base_bdevs": 4, 00:10:10.775 "num_base_bdevs_discovered": 2, 00:10:10.775 "num_base_bdevs_operational": 4, 00:10:10.775 "base_bdevs_list": [ 00:10:10.775 { 00:10:10.775 "name": "BaseBdev1", 00:10:10.775 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:10.775 "is_configured": true, 00:10:10.775 "data_offset": 2048, 00:10:10.775 "data_size": 63488 00:10:10.775 }, 00:10:10.775 { 00:10:10.775 "name": "BaseBdev2", 00:10:10.775 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:10.775 "is_configured": true, 00:10:10.775 "data_offset": 2048, 00:10:10.775 "data_size": 63488 00:10:10.775 }, 00:10:10.775 { 00:10:10.775 "name": "BaseBdev3", 00:10:10.775 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.775 "is_configured": false, 00:10:10.775 "data_offset": 0, 00:10:10.775 "data_size": 0 00:10:10.775 }, 00:10:10.775 { 00:10:10.775 "name": "BaseBdev4", 00:10:10.775 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.775 "is_configured": false, 00:10:10.775 "data_offset": 0, 00:10:10.775 "data_size": 0 00:10:10.775 } 00:10:10.775 ] 00:10:10.775 }' 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.775 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.036 [2024-10-29 10:59:16.475041] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:11.036 BaseBdev3 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.036 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.036 [ 00:10:11.036 { 00:10:11.036 "name": "BaseBdev3", 00:10:11.036 "aliases": [ 00:10:11.036 "52e1d15b-a482-4401-b053-a94b474892d0" 00:10:11.036 ], 00:10:11.036 "product_name": "Malloc disk", 00:10:11.036 "block_size": 512, 00:10:11.036 "num_blocks": 65536, 00:10:11.036 "uuid": "52e1d15b-a482-4401-b053-a94b474892d0", 00:10:11.036 "assigned_rate_limits": { 00:10:11.036 "rw_ios_per_sec": 0, 00:10:11.036 "rw_mbytes_per_sec": 0, 00:10:11.036 "r_mbytes_per_sec": 0, 00:10:11.036 "w_mbytes_per_sec": 0 00:10:11.036 }, 00:10:11.036 "claimed": true, 00:10:11.036 "claim_type": "exclusive_write", 00:10:11.036 "zoned": false, 00:10:11.036 "supported_io_types": { 00:10:11.036 "read": true, 00:10:11.036 "write": true, 00:10:11.036 "unmap": true, 00:10:11.036 "flush": true, 00:10:11.036 "reset": true, 00:10:11.036 "nvme_admin": false, 00:10:11.036 "nvme_io": false, 00:10:11.036 "nvme_io_md": false, 00:10:11.036 "write_zeroes": true, 00:10:11.036 "zcopy": true, 00:10:11.036 "get_zone_info": false, 00:10:11.036 "zone_management": false, 00:10:11.036 "zone_append": false, 00:10:11.036 "compare": false, 00:10:11.037 "compare_and_write": false, 00:10:11.037 "abort": true, 00:10:11.037 "seek_hole": false, 00:10:11.037 "seek_data": false, 00:10:11.037 "copy": true, 00:10:11.037 "nvme_iov_md": false 00:10:11.037 }, 00:10:11.037 "memory_domains": [ 00:10:11.037 { 00:10:11.037 "dma_device_id": "system", 00:10:11.037 "dma_device_type": 1 00:10:11.037 }, 00:10:11.037 { 00:10:11.037 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.037 "dma_device_type": 2 00:10:11.037 } 00:10:11.037 ], 00:10:11.037 "driver_specific": {} 00:10:11.037 } 00:10:11.037 ] 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.037 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.297 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.297 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.297 "name": "Existed_Raid", 00:10:11.297 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:11.297 "strip_size_kb": 64, 00:10:11.297 "state": "configuring", 00:10:11.297 "raid_level": "raid0", 00:10:11.297 "superblock": true, 00:10:11.297 "num_base_bdevs": 4, 00:10:11.297 "num_base_bdevs_discovered": 3, 00:10:11.297 "num_base_bdevs_operational": 4, 00:10:11.297 "base_bdevs_list": [ 00:10:11.297 { 00:10:11.297 "name": "BaseBdev1", 00:10:11.297 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:11.297 "is_configured": true, 00:10:11.297 "data_offset": 2048, 00:10:11.297 "data_size": 63488 00:10:11.297 }, 00:10:11.297 { 00:10:11.297 "name": "BaseBdev2", 00:10:11.297 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:11.297 "is_configured": true, 00:10:11.297 "data_offset": 2048, 00:10:11.297 "data_size": 63488 00:10:11.297 }, 00:10:11.297 { 00:10:11.297 "name": "BaseBdev3", 00:10:11.297 "uuid": "52e1d15b-a482-4401-b053-a94b474892d0", 00:10:11.297 "is_configured": true, 00:10:11.297 "data_offset": 2048, 00:10:11.297 "data_size": 63488 00:10:11.297 }, 00:10:11.297 { 00:10:11.297 "name": "BaseBdev4", 00:10:11.297 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.297 "is_configured": false, 00:10:11.297 "data_offset": 0, 00:10:11.297 "data_size": 0 00:10:11.297 } 00:10:11.297 ] 00:10:11.297 }' 00:10:11.297 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.297 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.558 [2024-10-29 10:59:16.975759] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:11.558 [2024-10-29 10:59:16.976161] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:11.558 [2024-10-29 10:59:16.976186] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:11.558 [2024-10-29 10:59:16.976552] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:11.558 BaseBdev4 00:10:11.558 [2024-10-29 10:59:16.976720] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:11.558 [2024-10-29 10:59:16.976769] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:11.558 [2024-10-29 10:59:16.976919] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.558 10:59:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.558 [ 00:10:11.558 { 00:10:11.558 "name": "BaseBdev4", 00:10:11.558 "aliases": [ 00:10:11.558 "d6c5615b-fbb0-4e09-a095-c5fb5e584582" 00:10:11.558 ], 00:10:11.558 "product_name": "Malloc disk", 00:10:11.558 "block_size": 512, 00:10:11.558 "num_blocks": 65536, 00:10:11.558 "uuid": "d6c5615b-fbb0-4e09-a095-c5fb5e584582", 00:10:11.558 "assigned_rate_limits": { 00:10:11.558 "rw_ios_per_sec": 0, 00:10:11.558 "rw_mbytes_per_sec": 0, 00:10:11.558 "r_mbytes_per_sec": 0, 00:10:11.558 "w_mbytes_per_sec": 0 00:10:11.558 }, 00:10:11.558 "claimed": true, 00:10:11.558 "claim_type": "exclusive_write", 00:10:11.558 "zoned": false, 00:10:11.558 "supported_io_types": { 00:10:11.558 "read": true, 00:10:11.558 "write": true, 00:10:11.558 "unmap": true, 00:10:11.558 "flush": true, 00:10:11.558 "reset": true, 00:10:11.558 "nvme_admin": false, 00:10:11.558 "nvme_io": false, 00:10:11.558 "nvme_io_md": false, 00:10:11.558 "write_zeroes": true, 00:10:11.558 "zcopy": true, 00:10:11.558 "get_zone_info": false, 00:10:11.558 "zone_management": false, 00:10:11.558 "zone_append": false, 00:10:11.558 "compare": false, 00:10:11.558 "compare_and_write": false, 00:10:11.558 "abort": true, 00:10:11.558 "seek_hole": false, 00:10:11.558 "seek_data": false, 00:10:11.558 "copy": true, 00:10:11.558 "nvme_iov_md": false 00:10:11.558 }, 00:10:11.558 "memory_domains": [ 00:10:11.558 { 00:10:11.558 "dma_device_id": "system", 00:10:11.558 "dma_device_type": 1 00:10:11.558 }, 00:10:11.558 { 00:10:11.558 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.558 "dma_device_type": 2 00:10:11.558 } 00:10:11.558 ], 00:10:11.558 "driver_specific": {} 00:10:11.558 } 00:10:11.558 ] 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.558 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.818 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.818 "name": "Existed_Raid", 00:10:11.818 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:11.818 "strip_size_kb": 64, 00:10:11.818 "state": "online", 00:10:11.818 "raid_level": "raid0", 00:10:11.818 "superblock": true, 00:10:11.818 "num_base_bdevs": 4, 00:10:11.818 "num_base_bdevs_discovered": 4, 00:10:11.818 "num_base_bdevs_operational": 4, 00:10:11.818 "base_bdevs_list": [ 00:10:11.818 { 00:10:11.818 "name": "BaseBdev1", 00:10:11.818 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:11.818 "is_configured": true, 00:10:11.818 "data_offset": 2048, 00:10:11.818 "data_size": 63488 00:10:11.818 }, 00:10:11.818 { 00:10:11.818 "name": "BaseBdev2", 00:10:11.818 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:11.818 "is_configured": true, 00:10:11.818 "data_offset": 2048, 00:10:11.818 "data_size": 63488 00:10:11.818 }, 00:10:11.818 { 00:10:11.818 "name": "BaseBdev3", 00:10:11.818 "uuid": "52e1d15b-a482-4401-b053-a94b474892d0", 00:10:11.818 "is_configured": true, 00:10:11.818 "data_offset": 2048, 00:10:11.818 "data_size": 63488 00:10:11.818 }, 00:10:11.818 { 00:10:11.818 "name": "BaseBdev4", 00:10:11.818 "uuid": "d6c5615b-fbb0-4e09-a095-c5fb5e584582", 00:10:11.818 "is_configured": true, 00:10:11.818 "data_offset": 2048, 00:10:11.818 "data_size": 63488 00:10:11.818 } 00:10:11.818 ] 00:10:11.818 }' 00:10:11.818 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.818 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.079 [2024-10-29 10:59:17.459706] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:12.079 "name": "Existed_Raid", 00:10:12.079 "aliases": [ 00:10:12.079 "dd1a9c70-d325-47cf-862b-454b030b564a" 00:10:12.079 ], 00:10:12.079 "product_name": "Raid Volume", 00:10:12.079 "block_size": 512, 00:10:12.079 "num_blocks": 253952, 00:10:12.079 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:12.079 "assigned_rate_limits": { 00:10:12.079 "rw_ios_per_sec": 0, 00:10:12.079 "rw_mbytes_per_sec": 0, 00:10:12.079 "r_mbytes_per_sec": 0, 00:10:12.079 "w_mbytes_per_sec": 0 00:10:12.079 }, 00:10:12.079 "claimed": false, 00:10:12.079 "zoned": false, 00:10:12.079 "supported_io_types": { 00:10:12.079 "read": true, 00:10:12.079 "write": true, 00:10:12.079 "unmap": true, 00:10:12.079 "flush": true, 00:10:12.079 "reset": true, 00:10:12.079 "nvme_admin": false, 00:10:12.079 "nvme_io": false, 00:10:12.079 "nvme_io_md": false, 00:10:12.079 "write_zeroes": true, 00:10:12.079 "zcopy": false, 00:10:12.079 "get_zone_info": false, 00:10:12.079 "zone_management": false, 00:10:12.079 "zone_append": false, 00:10:12.079 "compare": false, 00:10:12.079 "compare_and_write": false, 00:10:12.079 "abort": false, 00:10:12.079 "seek_hole": false, 00:10:12.079 "seek_data": false, 00:10:12.079 "copy": false, 00:10:12.079 "nvme_iov_md": false 00:10:12.079 }, 00:10:12.079 "memory_domains": [ 00:10:12.079 { 00:10:12.079 "dma_device_id": "system", 00:10:12.079 "dma_device_type": 1 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.079 "dma_device_type": 2 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "system", 00:10:12.079 "dma_device_type": 1 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.079 "dma_device_type": 2 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "system", 00:10:12.079 "dma_device_type": 1 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.079 "dma_device_type": 2 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "system", 00:10:12.079 "dma_device_type": 1 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:12.079 "dma_device_type": 2 00:10:12.079 } 00:10:12.079 ], 00:10:12.079 "driver_specific": { 00:10:12.079 "raid": { 00:10:12.079 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:12.079 "strip_size_kb": 64, 00:10:12.079 "state": "online", 00:10:12.079 "raid_level": "raid0", 00:10:12.079 "superblock": true, 00:10:12.079 "num_base_bdevs": 4, 00:10:12.079 "num_base_bdevs_discovered": 4, 00:10:12.079 "num_base_bdevs_operational": 4, 00:10:12.079 "base_bdevs_list": [ 00:10:12.079 { 00:10:12.079 "name": "BaseBdev1", 00:10:12.079 "uuid": "31cefd6c-deda-444c-986f-6ec8375d8abe", 00:10:12.079 "is_configured": true, 00:10:12.079 "data_offset": 2048, 00:10:12.079 "data_size": 63488 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "name": "BaseBdev2", 00:10:12.079 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:12.079 "is_configured": true, 00:10:12.079 "data_offset": 2048, 00:10:12.079 "data_size": 63488 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "name": "BaseBdev3", 00:10:12.079 "uuid": "52e1d15b-a482-4401-b053-a94b474892d0", 00:10:12.079 "is_configured": true, 00:10:12.079 "data_offset": 2048, 00:10:12.079 "data_size": 63488 00:10:12.079 }, 00:10:12.079 { 00:10:12.079 "name": "BaseBdev4", 00:10:12.079 "uuid": "d6c5615b-fbb0-4e09-a095-c5fb5e584582", 00:10:12.079 "is_configured": true, 00:10:12.079 "data_offset": 2048, 00:10:12.079 "data_size": 63488 00:10:12.079 } 00:10:12.079 ] 00:10:12.079 } 00:10:12.079 } 00:10:12.079 }' 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:12.079 BaseBdev2 00:10:12.079 BaseBdev3 00:10:12.079 BaseBdev4' 00:10:12.079 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.340 [2024-10-29 10:59:17.802765] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:12.340 [2024-10-29 10:59:17.802932] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:12.340 [2024-10-29 10:59:17.803024] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.340 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.600 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.600 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.600 "name": "Existed_Raid", 00:10:12.600 "uuid": "dd1a9c70-d325-47cf-862b-454b030b564a", 00:10:12.600 "strip_size_kb": 64, 00:10:12.600 "state": "offline", 00:10:12.600 "raid_level": "raid0", 00:10:12.600 "superblock": true, 00:10:12.600 "num_base_bdevs": 4, 00:10:12.600 "num_base_bdevs_discovered": 3, 00:10:12.600 "num_base_bdevs_operational": 3, 00:10:12.600 "base_bdevs_list": [ 00:10:12.600 { 00:10:12.600 "name": null, 00:10:12.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.600 "is_configured": false, 00:10:12.600 "data_offset": 0, 00:10:12.600 "data_size": 63488 00:10:12.600 }, 00:10:12.600 { 00:10:12.600 "name": "BaseBdev2", 00:10:12.600 "uuid": "14228ef6-78b2-43e5-9c88-0bd788d0ff86", 00:10:12.600 "is_configured": true, 00:10:12.600 "data_offset": 2048, 00:10:12.600 "data_size": 63488 00:10:12.600 }, 00:10:12.600 { 00:10:12.600 "name": "BaseBdev3", 00:10:12.600 "uuid": "52e1d15b-a482-4401-b053-a94b474892d0", 00:10:12.600 "is_configured": true, 00:10:12.600 "data_offset": 2048, 00:10:12.600 "data_size": 63488 00:10:12.600 }, 00:10:12.600 { 00:10:12.600 "name": "BaseBdev4", 00:10:12.600 "uuid": "d6c5615b-fbb0-4e09-a095-c5fb5e584582", 00:10:12.600 "is_configured": true, 00:10:12.600 "data_offset": 2048, 00:10:12.600 "data_size": 63488 00:10:12.600 } 00:10:12.600 ] 00:10:12.600 }' 00:10:12.600 10:59:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.600 10:59:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.861 [2024-10-29 10:59:18.303650] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.861 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.121 [2024-10-29 10:59:18.384671] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.121 [2024-10-29 10:59:18.457426] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:13.121 [2024-10-29 10:59:18.457582] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.121 BaseBdev2 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.121 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.122 [ 00:10:13.122 { 00:10:13.122 "name": "BaseBdev2", 00:10:13.122 "aliases": [ 00:10:13.122 "83149441-22a1-4d13-9b7f-f2de28d461f5" 00:10:13.122 ], 00:10:13.122 "product_name": "Malloc disk", 00:10:13.122 "block_size": 512, 00:10:13.122 "num_blocks": 65536, 00:10:13.122 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:13.122 "assigned_rate_limits": { 00:10:13.122 "rw_ios_per_sec": 0, 00:10:13.122 "rw_mbytes_per_sec": 0, 00:10:13.122 "r_mbytes_per_sec": 0, 00:10:13.122 "w_mbytes_per_sec": 0 00:10:13.122 }, 00:10:13.122 "claimed": false, 00:10:13.122 "zoned": false, 00:10:13.122 "supported_io_types": { 00:10:13.122 "read": true, 00:10:13.122 "write": true, 00:10:13.122 "unmap": true, 00:10:13.122 "flush": true, 00:10:13.122 "reset": true, 00:10:13.122 "nvme_admin": false, 00:10:13.122 "nvme_io": false, 00:10:13.122 "nvme_io_md": false, 00:10:13.122 "write_zeroes": true, 00:10:13.122 "zcopy": true, 00:10:13.122 "get_zone_info": false, 00:10:13.122 "zone_management": false, 00:10:13.122 "zone_append": false, 00:10:13.122 "compare": false, 00:10:13.122 "compare_and_write": false, 00:10:13.122 "abort": true, 00:10:13.122 "seek_hole": false, 00:10:13.122 "seek_data": false, 00:10:13.122 "copy": true, 00:10:13.122 "nvme_iov_md": false 00:10:13.122 }, 00:10:13.122 "memory_domains": [ 00:10:13.122 { 00:10:13.122 "dma_device_id": "system", 00:10:13.122 "dma_device_type": 1 00:10:13.122 }, 00:10:13.122 { 00:10:13.122 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.122 "dma_device_type": 2 00:10:13.122 } 00:10:13.122 ], 00:10:13.122 "driver_specific": {} 00:10:13.122 } 00:10:13.122 ] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.122 BaseBdev3 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.122 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.382 [ 00:10:13.382 { 00:10:13.382 "name": "BaseBdev3", 00:10:13.382 "aliases": [ 00:10:13.382 "b7ffae4e-0739-4c66-8248-bce4c721e53d" 00:10:13.382 ], 00:10:13.382 "product_name": "Malloc disk", 00:10:13.382 "block_size": 512, 00:10:13.382 "num_blocks": 65536, 00:10:13.382 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:13.382 "assigned_rate_limits": { 00:10:13.382 "rw_ios_per_sec": 0, 00:10:13.382 "rw_mbytes_per_sec": 0, 00:10:13.382 "r_mbytes_per_sec": 0, 00:10:13.382 "w_mbytes_per_sec": 0 00:10:13.382 }, 00:10:13.382 "claimed": false, 00:10:13.382 "zoned": false, 00:10:13.382 "supported_io_types": { 00:10:13.382 "read": true, 00:10:13.382 "write": true, 00:10:13.382 "unmap": true, 00:10:13.382 "flush": true, 00:10:13.382 "reset": true, 00:10:13.382 "nvme_admin": false, 00:10:13.382 "nvme_io": false, 00:10:13.382 "nvme_io_md": false, 00:10:13.382 "write_zeroes": true, 00:10:13.382 "zcopy": true, 00:10:13.382 "get_zone_info": false, 00:10:13.382 "zone_management": false, 00:10:13.382 "zone_append": false, 00:10:13.382 "compare": false, 00:10:13.382 "compare_and_write": false, 00:10:13.382 "abort": true, 00:10:13.382 "seek_hole": false, 00:10:13.382 "seek_data": false, 00:10:13.382 "copy": true, 00:10:13.382 "nvme_iov_md": false 00:10:13.382 }, 00:10:13.382 "memory_domains": [ 00:10:13.382 { 00:10:13.382 "dma_device_id": "system", 00:10:13.382 "dma_device_type": 1 00:10:13.382 }, 00:10:13.382 { 00:10:13.382 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.382 "dma_device_type": 2 00:10:13.382 } 00:10:13.382 ], 00:10:13.382 "driver_specific": {} 00:10:13.382 } 00:10:13.382 ] 00:10:13.382 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.382 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:13.382 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.383 BaseBdev4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.383 [ 00:10:13.383 { 00:10:13.383 "name": "BaseBdev4", 00:10:13.383 "aliases": [ 00:10:13.383 "91575b46-d214-4cee-a210-ea67a12fd180" 00:10:13.383 ], 00:10:13.383 "product_name": "Malloc disk", 00:10:13.383 "block_size": 512, 00:10:13.383 "num_blocks": 65536, 00:10:13.383 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:13.383 "assigned_rate_limits": { 00:10:13.383 "rw_ios_per_sec": 0, 00:10:13.383 "rw_mbytes_per_sec": 0, 00:10:13.383 "r_mbytes_per_sec": 0, 00:10:13.383 "w_mbytes_per_sec": 0 00:10:13.383 }, 00:10:13.383 "claimed": false, 00:10:13.383 "zoned": false, 00:10:13.383 "supported_io_types": { 00:10:13.383 "read": true, 00:10:13.383 "write": true, 00:10:13.383 "unmap": true, 00:10:13.383 "flush": true, 00:10:13.383 "reset": true, 00:10:13.383 "nvme_admin": false, 00:10:13.383 "nvme_io": false, 00:10:13.383 "nvme_io_md": false, 00:10:13.383 "write_zeroes": true, 00:10:13.383 "zcopy": true, 00:10:13.383 "get_zone_info": false, 00:10:13.383 "zone_management": false, 00:10:13.383 "zone_append": false, 00:10:13.383 "compare": false, 00:10:13.383 "compare_and_write": false, 00:10:13.383 "abort": true, 00:10:13.383 "seek_hole": false, 00:10:13.383 "seek_data": false, 00:10:13.383 "copy": true, 00:10:13.383 "nvme_iov_md": false 00:10:13.383 }, 00:10:13.383 "memory_domains": [ 00:10:13.383 { 00:10:13.383 "dma_device_id": "system", 00:10:13.383 "dma_device_type": 1 00:10:13.383 }, 00:10:13.383 { 00:10:13.383 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.383 "dma_device_type": 2 00:10:13.383 } 00:10:13.383 ], 00:10:13.383 "driver_specific": {} 00:10:13.383 } 00:10:13.383 ] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.383 [2024-10-29 10:59:18.707917] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:13.383 [2024-10-29 10:59:18.708052] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:13.383 [2024-10-29 10:59:18.708105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:13.383 [2024-10-29 10:59:18.710454] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:13.383 [2024-10-29 10:59:18.710556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.383 "name": "Existed_Raid", 00:10:13.383 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:13.383 "strip_size_kb": 64, 00:10:13.383 "state": "configuring", 00:10:13.383 "raid_level": "raid0", 00:10:13.383 "superblock": true, 00:10:13.383 "num_base_bdevs": 4, 00:10:13.383 "num_base_bdevs_discovered": 3, 00:10:13.383 "num_base_bdevs_operational": 4, 00:10:13.383 "base_bdevs_list": [ 00:10:13.383 { 00:10:13.383 "name": "BaseBdev1", 00:10:13.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.383 "is_configured": false, 00:10:13.383 "data_offset": 0, 00:10:13.383 "data_size": 0 00:10:13.383 }, 00:10:13.383 { 00:10:13.383 "name": "BaseBdev2", 00:10:13.383 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:13.383 "is_configured": true, 00:10:13.383 "data_offset": 2048, 00:10:13.383 "data_size": 63488 00:10:13.383 }, 00:10:13.383 { 00:10:13.383 "name": "BaseBdev3", 00:10:13.383 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:13.383 "is_configured": true, 00:10:13.383 "data_offset": 2048, 00:10:13.383 "data_size": 63488 00:10:13.383 }, 00:10:13.383 { 00:10:13.383 "name": "BaseBdev4", 00:10:13.383 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:13.383 "is_configured": true, 00:10:13.383 "data_offset": 2048, 00:10:13.383 "data_size": 63488 00:10:13.383 } 00:10:13.383 ] 00:10:13.383 }' 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.383 10:59:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.954 [2024-10-29 10:59:19.155594] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.954 "name": "Existed_Raid", 00:10:13.954 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:13.954 "strip_size_kb": 64, 00:10:13.954 "state": "configuring", 00:10:13.954 "raid_level": "raid0", 00:10:13.954 "superblock": true, 00:10:13.954 "num_base_bdevs": 4, 00:10:13.954 "num_base_bdevs_discovered": 2, 00:10:13.954 "num_base_bdevs_operational": 4, 00:10:13.954 "base_bdevs_list": [ 00:10:13.954 { 00:10:13.954 "name": "BaseBdev1", 00:10:13.954 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.954 "is_configured": false, 00:10:13.954 "data_offset": 0, 00:10:13.954 "data_size": 0 00:10:13.954 }, 00:10:13.954 { 00:10:13.954 "name": null, 00:10:13.954 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:13.954 "is_configured": false, 00:10:13.954 "data_offset": 0, 00:10:13.954 "data_size": 63488 00:10:13.954 }, 00:10:13.954 { 00:10:13.954 "name": "BaseBdev3", 00:10:13.954 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:13.954 "is_configured": true, 00:10:13.954 "data_offset": 2048, 00:10:13.954 "data_size": 63488 00:10:13.954 }, 00:10:13.954 { 00:10:13.954 "name": "BaseBdev4", 00:10:13.954 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:13.954 "is_configured": true, 00:10:13.954 "data_offset": 2048, 00:10:13.954 "data_size": 63488 00:10:13.954 } 00:10:13.954 ] 00:10:13.954 }' 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.954 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.215 [2024-10-29 10:59:19.584658] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:14.215 BaseBdev1 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:14.215 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.216 [ 00:10:14.216 { 00:10:14.216 "name": "BaseBdev1", 00:10:14.216 "aliases": [ 00:10:14.216 "ef407af7-fd7e-47fa-8da1-7e829620f155" 00:10:14.216 ], 00:10:14.216 "product_name": "Malloc disk", 00:10:14.216 "block_size": 512, 00:10:14.216 "num_blocks": 65536, 00:10:14.216 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:14.216 "assigned_rate_limits": { 00:10:14.216 "rw_ios_per_sec": 0, 00:10:14.216 "rw_mbytes_per_sec": 0, 00:10:14.216 "r_mbytes_per_sec": 0, 00:10:14.216 "w_mbytes_per_sec": 0 00:10:14.216 }, 00:10:14.216 "claimed": true, 00:10:14.216 "claim_type": "exclusive_write", 00:10:14.216 "zoned": false, 00:10:14.216 "supported_io_types": { 00:10:14.216 "read": true, 00:10:14.216 "write": true, 00:10:14.216 "unmap": true, 00:10:14.216 "flush": true, 00:10:14.216 "reset": true, 00:10:14.216 "nvme_admin": false, 00:10:14.216 "nvme_io": false, 00:10:14.216 "nvme_io_md": false, 00:10:14.216 "write_zeroes": true, 00:10:14.216 "zcopy": true, 00:10:14.216 "get_zone_info": false, 00:10:14.216 "zone_management": false, 00:10:14.216 "zone_append": false, 00:10:14.216 "compare": false, 00:10:14.216 "compare_and_write": false, 00:10:14.216 "abort": true, 00:10:14.216 "seek_hole": false, 00:10:14.216 "seek_data": false, 00:10:14.216 "copy": true, 00:10:14.216 "nvme_iov_md": false 00:10:14.216 }, 00:10:14.216 "memory_domains": [ 00:10:14.216 { 00:10:14.216 "dma_device_id": "system", 00:10:14.216 "dma_device_type": 1 00:10:14.216 }, 00:10:14.216 { 00:10:14.216 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.216 "dma_device_type": 2 00:10:14.216 } 00:10:14.216 ], 00:10:14.216 "driver_specific": {} 00:10:14.216 } 00:10:14.216 ] 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.216 "name": "Existed_Raid", 00:10:14.216 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:14.216 "strip_size_kb": 64, 00:10:14.216 "state": "configuring", 00:10:14.216 "raid_level": "raid0", 00:10:14.216 "superblock": true, 00:10:14.216 "num_base_bdevs": 4, 00:10:14.216 "num_base_bdevs_discovered": 3, 00:10:14.216 "num_base_bdevs_operational": 4, 00:10:14.216 "base_bdevs_list": [ 00:10:14.216 { 00:10:14.216 "name": "BaseBdev1", 00:10:14.216 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:14.216 "is_configured": true, 00:10:14.216 "data_offset": 2048, 00:10:14.216 "data_size": 63488 00:10:14.216 }, 00:10:14.216 { 00:10:14.216 "name": null, 00:10:14.216 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:14.216 "is_configured": false, 00:10:14.216 "data_offset": 0, 00:10:14.216 "data_size": 63488 00:10:14.216 }, 00:10:14.216 { 00:10:14.216 "name": "BaseBdev3", 00:10:14.216 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:14.216 "is_configured": true, 00:10:14.216 "data_offset": 2048, 00:10:14.216 "data_size": 63488 00:10:14.216 }, 00:10:14.216 { 00:10:14.216 "name": "BaseBdev4", 00:10:14.216 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:14.216 "is_configured": true, 00:10:14.216 "data_offset": 2048, 00:10:14.216 "data_size": 63488 00:10:14.216 } 00:10:14.216 ] 00:10:14.216 }' 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.216 10:59:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.784 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.784 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.784 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.785 [2024-10-29 10:59:20.147826] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.785 "name": "Existed_Raid", 00:10:14.785 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:14.785 "strip_size_kb": 64, 00:10:14.785 "state": "configuring", 00:10:14.785 "raid_level": "raid0", 00:10:14.785 "superblock": true, 00:10:14.785 "num_base_bdevs": 4, 00:10:14.785 "num_base_bdevs_discovered": 2, 00:10:14.785 "num_base_bdevs_operational": 4, 00:10:14.785 "base_bdevs_list": [ 00:10:14.785 { 00:10:14.785 "name": "BaseBdev1", 00:10:14.785 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:14.785 "is_configured": true, 00:10:14.785 "data_offset": 2048, 00:10:14.785 "data_size": 63488 00:10:14.785 }, 00:10:14.785 { 00:10:14.785 "name": null, 00:10:14.785 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:14.785 "is_configured": false, 00:10:14.785 "data_offset": 0, 00:10:14.785 "data_size": 63488 00:10:14.785 }, 00:10:14.785 { 00:10:14.785 "name": null, 00:10:14.785 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:14.785 "is_configured": false, 00:10:14.785 "data_offset": 0, 00:10:14.785 "data_size": 63488 00:10:14.785 }, 00:10:14.785 { 00:10:14.785 "name": "BaseBdev4", 00:10:14.785 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:14.785 "is_configured": true, 00:10:14.785 "data_offset": 2048, 00:10:14.785 "data_size": 63488 00:10:14.785 } 00:10:14.785 ] 00:10:14.785 }' 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.785 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.353 [2024-10-29 10:59:20.643632] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.353 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.353 "name": "Existed_Raid", 00:10:15.353 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:15.353 "strip_size_kb": 64, 00:10:15.353 "state": "configuring", 00:10:15.353 "raid_level": "raid0", 00:10:15.353 "superblock": true, 00:10:15.353 "num_base_bdevs": 4, 00:10:15.353 "num_base_bdevs_discovered": 3, 00:10:15.353 "num_base_bdevs_operational": 4, 00:10:15.353 "base_bdevs_list": [ 00:10:15.353 { 00:10:15.353 "name": "BaseBdev1", 00:10:15.353 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:15.353 "is_configured": true, 00:10:15.353 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": null, 00:10:15.354 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:15.354 "is_configured": false, 00:10:15.354 "data_offset": 0, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": "BaseBdev3", 00:10:15.354 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": "BaseBdev4", 00:10:15.354 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 } 00:10:15.354 ] 00:10:15.354 }' 00:10:15.354 10:59:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.354 10:59:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 [2024-10-29 10:59:21.167630] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.923 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.923 "name": "Existed_Raid", 00:10:15.923 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:15.923 "strip_size_kb": 64, 00:10:15.923 "state": "configuring", 00:10:15.923 "raid_level": "raid0", 00:10:15.923 "superblock": true, 00:10:15.923 "num_base_bdevs": 4, 00:10:15.923 "num_base_bdevs_discovered": 2, 00:10:15.923 "num_base_bdevs_operational": 4, 00:10:15.923 "base_bdevs_list": [ 00:10:15.923 { 00:10:15.923 "name": null, 00:10:15.923 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:15.923 "is_configured": false, 00:10:15.923 "data_offset": 0, 00:10:15.923 "data_size": 63488 00:10:15.923 }, 00:10:15.923 { 00:10:15.923 "name": null, 00:10:15.923 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:15.924 "is_configured": false, 00:10:15.924 "data_offset": 0, 00:10:15.924 "data_size": 63488 00:10:15.924 }, 00:10:15.924 { 00:10:15.924 "name": "BaseBdev3", 00:10:15.924 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:15.924 "is_configured": true, 00:10:15.924 "data_offset": 2048, 00:10:15.924 "data_size": 63488 00:10:15.924 }, 00:10:15.924 { 00:10:15.924 "name": "BaseBdev4", 00:10:15.924 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:15.924 "is_configured": true, 00:10:15.924 "data_offset": 2048, 00:10:15.924 "data_size": 63488 00:10:15.924 } 00:10:15.924 ] 00:10:15.924 }' 00:10:15.924 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.924 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:16.183 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.184 [2024-10-29 10:59:21.631584] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.184 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.443 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.443 "name": "Existed_Raid", 00:10:16.443 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:16.443 "strip_size_kb": 64, 00:10:16.443 "state": "configuring", 00:10:16.443 "raid_level": "raid0", 00:10:16.443 "superblock": true, 00:10:16.443 "num_base_bdevs": 4, 00:10:16.443 "num_base_bdevs_discovered": 3, 00:10:16.443 "num_base_bdevs_operational": 4, 00:10:16.443 "base_bdevs_list": [ 00:10:16.443 { 00:10:16.443 "name": null, 00:10:16.443 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:16.443 "is_configured": false, 00:10:16.443 "data_offset": 0, 00:10:16.443 "data_size": 63488 00:10:16.443 }, 00:10:16.443 { 00:10:16.443 "name": "BaseBdev2", 00:10:16.443 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:16.443 "is_configured": true, 00:10:16.443 "data_offset": 2048, 00:10:16.443 "data_size": 63488 00:10:16.443 }, 00:10:16.443 { 00:10:16.443 "name": "BaseBdev3", 00:10:16.443 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:16.443 "is_configured": true, 00:10:16.443 "data_offset": 2048, 00:10:16.443 "data_size": 63488 00:10:16.443 }, 00:10:16.443 { 00:10:16.443 "name": "BaseBdev4", 00:10:16.443 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:16.443 "is_configured": true, 00:10:16.443 "data_offset": 2048, 00:10:16.443 "data_size": 63488 00:10:16.443 } 00:10:16.443 ] 00:10:16.443 }' 00:10:16.443 10:59:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.443 10:59:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:16.702 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.703 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ef407af7-fd7e-47fa-8da1-7e829620f155 00:10:16.703 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.703 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.962 [2024-10-29 10:59:22.208934] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:16.962 [2024-10-29 10:59:22.209305] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:16.962 [2024-10-29 10:59:22.209367] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:16.962 [2024-10-29 10:59:22.209746] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:16.962 [2024-10-29 10:59:22.209926] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:16.962 NewBaseBdev 00:10:16.962 [2024-10-29 10:59:22.209978] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:16.962 [2024-10-29 10:59:22.210168] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.962 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.962 [ 00:10:16.962 { 00:10:16.962 "name": "NewBaseBdev", 00:10:16.962 "aliases": [ 00:10:16.962 "ef407af7-fd7e-47fa-8da1-7e829620f155" 00:10:16.962 ], 00:10:16.962 "product_name": "Malloc disk", 00:10:16.962 "block_size": 512, 00:10:16.962 "num_blocks": 65536, 00:10:16.962 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:16.962 "assigned_rate_limits": { 00:10:16.962 "rw_ios_per_sec": 0, 00:10:16.962 "rw_mbytes_per_sec": 0, 00:10:16.962 "r_mbytes_per_sec": 0, 00:10:16.962 "w_mbytes_per_sec": 0 00:10:16.962 }, 00:10:16.962 "claimed": true, 00:10:16.962 "claim_type": "exclusive_write", 00:10:16.962 "zoned": false, 00:10:16.962 "supported_io_types": { 00:10:16.962 "read": true, 00:10:16.962 "write": true, 00:10:16.962 "unmap": true, 00:10:16.962 "flush": true, 00:10:16.962 "reset": true, 00:10:16.962 "nvme_admin": false, 00:10:16.962 "nvme_io": false, 00:10:16.962 "nvme_io_md": false, 00:10:16.962 "write_zeroes": true, 00:10:16.962 "zcopy": true, 00:10:16.963 "get_zone_info": false, 00:10:16.963 "zone_management": false, 00:10:16.963 "zone_append": false, 00:10:16.963 "compare": false, 00:10:16.963 "compare_and_write": false, 00:10:16.963 "abort": true, 00:10:16.963 "seek_hole": false, 00:10:16.963 "seek_data": false, 00:10:16.963 "copy": true, 00:10:16.963 "nvme_iov_md": false 00:10:16.963 }, 00:10:16.963 "memory_domains": [ 00:10:16.963 { 00:10:16.963 "dma_device_id": "system", 00:10:16.963 "dma_device_type": 1 00:10:16.963 }, 00:10:16.963 { 00:10:16.963 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.963 "dma_device_type": 2 00:10:16.963 } 00:10:16.963 ], 00:10:16.963 "driver_specific": {} 00:10:16.963 } 00:10:16.963 ] 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.963 "name": "Existed_Raid", 00:10:16.963 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:16.963 "strip_size_kb": 64, 00:10:16.963 "state": "online", 00:10:16.963 "raid_level": "raid0", 00:10:16.963 "superblock": true, 00:10:16.963 "num_base_bdevs": 4, 00:10:16.963 "num_base_bdevs_discovered": 4, 00:10:16.963 "num_base_bdevs_operational": 4, 00:10:16.963 "base_bdevs_list": [ 00:10:16.963 { 00:10:16.963 "name": "NewBaseBdev", 00:10:16.963 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:16.963 "is_configured": true, 00:10:16.963 "data_offset": 2048, 00:10:16.963 "data_size": 63488 00:10:16.963 }, 00:10:16.963 { 00:10:16.963 "name": "BaseBdev2", 00:10:16.963 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:16.963 "is_configured": true, 00:10:16.963 "data_offset": 2048, 00:10:16.963 "data_size": 63488 00:10:16.963 }, 00:10:16.963 { 00:10:16.963 "name": "BaseBdev3", 00:10:16.963 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:16.963 "is_configured": true, 00:10:16.963 "data_offset": 2048, 00:10:16.963 "data_size": 63488 00:10:16.963 }, 00:10:16.963 { 00:10:16.963 "name": "BaseBdev4", 00:10:16.963 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:16.963 "is_configured": true, 00:10:16.963 "data_offset": 2048, 00:10:16.963 "data_size": 63488 00:10:16.963 } 00:10:16.963 ] 00:10:16.963 }' 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.963 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:17.222 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.222 [2024-10-29 10:59:22.700650] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:17.482 "name": "Existed_Raid", 00:10:17.482 "aliases": [ 00:10:17.482 "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5" 00:10:17.482 ], 00:10:17.482 "product_name": "Raid Volume", 00:10:17.482 "block_size": 512, 00:10:17.482 "num_blocks": 253952, 00:10:17.482 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:17.482 "assigned_rate_limits": { 00:10:17.482 "rw_ios_per_sec": 0, 00:10:17.482 "rw_mbytes_per_sec": 0, 00:10:17.482 "r_mbytes_per_sec": 0, 00:10:17.482 "w_mbytes_per_sec": 0 00:10:17.482 }, 00:10:17.482 "claimed": false, 00:10:17.482 "zoned": false, 00:10:17.482 "supported_io_types": { 00:10:17.482 "read": true, 00:10:17.482 "write": true, 00:10:17.482 "unmap": true, 00:10:17.482 "flush": true, 00:10:17.482 "reset": true, 00:10:17.482 "nvme_admin": false, 00:10:17.482 "nvme_io": false, 00:10:17.482 "nvme_io_md": false, 00:10:17.482 "write_zeroes": true, 00:10:17.482 "zcopy": false, 00:10:17.482 "get_zone_info": false, 00:10:17.482 "zone_management": false, 00:10:17.482 "zone_append": false, 00:10:17.482 "compare": false, 00:10:17.482 "compare_and_write": false, 00:10:17.482 "abort": false, 00:10:17.482 "seek_hole": false, 00:10:17.482 "seek_data": false, 00:10:17.482 "copy": false, 00:10:17.482 "nvme_iov_md": false 00:10:17.482 }, 00:10:17.482 "memory_domains": [ 00:10:17.482 { 00:10:17.482 "dma_device_id": "system", 00:10:17.482 "dma_device_type": 1 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.482 "dma_device_type": 2 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "system", 00:10:17.482 "dma_device_type": 1 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.482 "dma_device_type": 2 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "system", 00:10:17.482 "dma_device_type": 1 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.482 "dma_device_type": 2 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "system", 00:10:17.482 "dma_device_type": 1 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.482 "dma_device_type": 2 00:10:17.482 } 00:10:17.482 ], 00:10:17.482 "driver_specific": { 00:10:17.482 "raid": { 00:10:17.482 "uuid": "f6763a6d-6fe1-4f57-b43c-96bc0bdfb8c5", 00:10:17.482 "strip_size_kb": 64, 00:10:17.482 "state": "online", 00:10:17.482 "raid_level": "raid0", 00:10:17.482 "superblock": true, 00:10:17.482 "num_base_bdevs": 4, 00:10:17.482 "num_base_bdevs_discovered": 4, 00:10:17.482 "num_base_bdevs_operational": 4, 00:10:17.482 "base_bdevs_list": [ 00:10:17.482 { 00:10:17.482 "name": "NewBaseBdev", 00:10:17.482 "uuid": "ef407af7-fd7e-47fa-8da1-7e829620f155", 00:10:17.482 "is_configured": true, 00:10:17.482 "data_offset": 2048, 00:10:17.482 "data_size": 63488 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "name": "BaseBdev2", 00:10:17.482 "uuid": "83149441-22a1-4d13-9b7f-f2de28d461f5", 00:10:17.482 "is_configured": true, 00:10:17.482 "data_offset": 2048, 00:10:17.482 "data_size": 63488 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "name": "BaseBdev3", 00:10:17.482 "uuid": "b7ffae4e-0739-4c66-8248-bce4c721e53d", 00:10:17.482 "is_configured": true, 00:10:17.482 "data_offset": 2048, 00:10:17.482 "data_size": 63488 00:10:17.482 }, 00:10:17.482 { 00:10:17.482 "name": "BaseBdev4", 00:10:17.482 "uuid": "91575b46-d214-4cee-a210-ea67a12fd180", 00:10:17.482 "is_configured": true, 00:10:17.482 "data_offset": 2048, 00:10:17.482 "data_size": 63488 00:10:17.482 } 00:10:17.482 ] 00:10:17.482 } 00:10:17.482 } 00:10:17.482 }' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:17.482 BaseBdev2 00:10:17.482 BaseBdev3 00:10:17.482 BaseBdev4' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:17.482 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:17.483 10:59:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:17.483 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.483 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.742 10:59:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.742 [2024-10-29 10:59:23.023674] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:17.742 [2024-10-29 10:59:23.023800] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:17.742 [2024-10-29 10:59:23.023951] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:17.742 [2024-10-29 10:59:23.024057] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:17.742 [2024-10-29 10:59:23.024070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81159 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 81159 ']' 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 81159 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 81159 00:10:17.742 killing process with pid 81159 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:17.742 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 81159' 00:10:17.743 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 81159 00:10:17.743 [2024-10-29 10:59:23.071085] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:17.743 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 81159 00:10:17.743 [2024-10-29 10:59:23.151405] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:18.003 10:59:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:18.003 00:10:18.003 real 0m9.951s 00:10:18.003 user 0m16.687s 00:10:18.003 sys 0m2.125s 00:10:18.003 ************************************ 00:10:18.003 END TEST raid_state_function_test_sb 00:10:18.003 ************************************ 00:10:18.003 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:18.003 10:59:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.264 10:59:23 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:10:18.264 10:59:23 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:10:18.264 10:59:23 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:18.264 10:59:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:18.264 ************************************ 00:10:18.264 START TEST raid_superblock_test 00:10:18.264 ************************************ 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid0 4 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81807 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81807 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 81807 ']' 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:18.264 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:18.264 10:59:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.264 [2024-10-29 10:59:23.645609] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:18.264 [2024-10-29 10:59:23.645836] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81807 ] 00:10:18.524 [2024-10-29 10:59:23.817812] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:18.525 [2024-10-29 10:59:23.861768] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:18.525 [2024-10-29 10:59:23.941036] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:18.525 [2024-10-29 10:59:23.941193] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:19.096 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 malloc1 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 [2024-10-29 10:59:24.513915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:19.097 [2024-10-29 10:59:24.514082] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.097 [2024-10-29 10:59:24.514130] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:19.097 [2024-10-29 10:59:24.514182] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.097 [2024-10-29 10:59:24.516840] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.097 [2024-10-29 10:59:24.516926] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:19.097 pt1 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 malloc2 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 [2024-10-29 10:59:24.553314] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:19.097 [2024-10-29 10:59:24.553401] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.097 [2024-10-29 10:59:24.553422] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:19.097 [2024-10-29 10:59:24.553435] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.097 [2024-10-29 10:59:24.556089] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.097 [2024-10-29 10:59:24.556219] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:19.097 pt2 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 malloc3 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.097 [2024-10-29 10:59:24.588402] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:19.097 [2024-10-29 10:59:24.588541] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.097 [2024-10-29 10:59:24.588582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:19.097 [2024-10-29 10:59:24.588616] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.097 [2024-10-29 10:59:24.591288] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.097 [2024-10-29 10:59:24.591399] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:19.097 pt3 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:19.097 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.359 malloc4 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.359 [2024-10-29 10:59:24.639006] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:19.359 [2024-10-29 10:59:24.639142] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.359 [2024-10-29 10:59:24.639180] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:19.359 [2024-10-29 10:59:24.639216] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.359 [2024-10-29 10:59:24.641737] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.359 [2024-10-29 10:59:24.641815] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:19.359 pt4 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.359 [2024-10-29 10:59:24.651065] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:19.359 [2024-10-29 10:59:24.653313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:19.359 [2024-10-29 10:59:24.653438] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:19.359 [2024-10-29 10:59:24.653528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:19.359 [2024-10-29 10:59:24.653779] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:10:19.359 [2024-10-29 10:59:24.653831] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:19.359 [2024-10-29 10:59:24.654181] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:19.359 [2024-10-29 10:59:24.654421] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:10:19.359 [2024-10-29 10:59:24.654467] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:10:19.359 [2024-10-29 10:59:24.654703] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.359 "name": "raid_bdev1", 00:10:19.359 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:19.359 "strip_size_kb": 64, 00:10:19.359 "state": "online", 00:10:19.359 "raid_level": "raid0", 00:10:19.359 "superblock": true, 00:10:19.359 "num_base_bdevs": 4, 00:10:19.359 "num_base_bdevs_discovered": 4, 00:10:19.359 "num_base_bdevs_operational": 4, 00:10:19.359 "base_bdevs_list": [ 00:10:19.359 { 00:10:19.359 "name": "pt1", 00:10:19.359 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:19.359 "is_configured": true, 00:10:19.359 "data_offset": 2048, 00:10:19.359 "data_size": 63488 00:10:19.359 }, 00:10:19.359 { 00:10:19.359 "name": "pt2", 00:10:19.359 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:19.359 "is_configured": true, 00:10:19.359 "data_offset": 2048, 00:10:19.359 "data_size": 63488 00:10:19.359 }, 00:10:19.359 { 00:10:19.359 "name": "pt3", 00:10:19.359 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:19.359 "is_configured": true, 00:10:19.359 "data_offset": 2048, 00:10:19.359 "data_size": 63488 00:10:19.359 }, 00:10:19.359 { 00:10:19.359 "name": "pt4", 00:10:19.359 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:19.359 "is_configured": true, 00:10:19.359 "data_offset": 2048, 00:10:19.359 "data_size": 63488 00:10:19.359 } 00:10:19.359 ] 00:10:19.359 }' 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.359 10:59:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.620 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:19.620 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.882 [2024-10-29 10:59:25.134656] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:19.882 "name": "raid_bdev1", 00:10:19.882 "aliases": [ 00:10:19.882 "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730" 00:10:19.882 ], 00:10:19.882 "product_name": "Raid Volume", 00:10:19.882 "block_size": 512, 00:10:19.882 "num_blocks": 253952, 00:10:19.882 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:19.882 "assigned_rate_limits": { 00:10:19.882 "rw_ios_per_sec": 0, 00:10:19.882 "rw_mbytes_per_sec": 0, 00:10:19.882 "r_mbytes_per_sec": 0, 00:10:19.882 "w_mbytes_per_sec": 0 00:10:19.882 }, 00:10:19.882 "claimed": false, 00:10:19.882 "zoned": false, 00:10:19.882 "supported_io_types": { 00:10:19.882 "read": true, 00:10:19.882 "write": true, 00:10:19.882 "unmap": true, 00:10:19.882 "flush": true, 00:10:19.882 "reset": true, 00:10:19.882 "nvme_admin": false, 00:10:19.882 "nvme_io": false, 00:10:19.882 "nvme_io_md": false, 00:10:19.882 "write_zeroes": true, 00:10:19.882 "zcopy": false, 00:10:19.882 "get_zone_info": false, 00:10:19.882 "zone_management": false, 00:10:19.882 "zone_append": false, 00:10:19.882 "compare": false, 00:10:19.882 "compare_and_write": false, 00:10:19.882 "abort": false, 00:10:19.882 "seek_hole": false, 00:10:19.882 "seek_data": false, 00:10:19.882 "copy": false, 00:10:19.882 "nvme_iov_md": false 00:10:19.882 }, 00:10:19.882 "memory_domains": [ 00:10:19.882 { 00:10:19.882 "dma_device_id": "system", 00:10:19.882 "dma_device_type": 1 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.882 "dma_device_type": 2 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "system", 00:10:19.882 "dma_device_type": 1 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.882 "dma_device_type": 2 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "system", 00:10:19.882 "dma_device_type": 1 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.882 "dma_device_type": 2 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "system", 00:10:19.882 "dma_device_type": 1 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.882 "dma_device_type": 2 00:10:19.882 } 00:10:19.882 ], 00:10:19.882 "driver_specific": { 00:10:19.882 "raid": { 00:10:19.882 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:19.882 "strip_size_kb": 64, 00:10:19.882 "state": "online", 00:10:19.882 "raid_level": "raid0", 00:10:19.882 "superblock": true, 00:10:19.882 "num_base_bdevs": 4, 00:10:19.882 "num_base_bdevs_discovered": 4, 00:10:19.882 "num_base_bdevs_operational": 4, 00:10:19.882 "base_bdevs_list": [ 00:10:19.882 { 00:10:19.882 "name": "pt1", 00:10:19.882 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:19.882 "is_configured": true, 00:10:19.882 "data_offset": 2048, 00:10:19.882 "data_size": 63488 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "name": "pt2", 00:10:19.882 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:19.882 "is_configured": true, 00:10:19.882 "data_offset": 2048, 00:10:19.882 "data_size": 63488 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "name": "pt3", 00:10:19.882 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:19.882 "is_configured": true, 00:10:19.882 "data_offset": 2048, 00:10:19.882 "data_size": 63488 00:10:19.882 }, 00:10:19.882 { 00:10:19.882 "name": "pt4", 00:10:19.882 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:19.882 "is_configured": true, 00:10:19.882 "data_offset": 2048, 00:10:19.882 "data_size": 63488 00:10:19.882 } 00:10:19.882 ] 00:10:19.882 } 00:10:19.882 } 00:10:19.882 }' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:19.882 pt2 00:10:19.882 pt3 00:10:19.882 pt4' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:19.882 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.143 [2024-10-29 10:59:25.454015] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bd5cd3c1-861a-45d4-afbe-5c1fcdde3730 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z bd5cd3c1-861a-45d4-afbe-5c1fcdde3730 ']' 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.143 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.143 [2024-10-29 10:59:25.501578] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:20.144 [2024-10-29 10:59:25.501708] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:20.144 [2024-10-29 10:59:25.501842] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:20.144 [2024-10-29 10:59:25.501973] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:20.144 [2024-10-29 10:59:25.502038] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:20.144 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.405 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.405 [2024-10-29 10:59:25.665563] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:20.405 [2024-10-29 10:59:25.668020] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:20.405 [2024-10-29 10:59:25.668080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:20.405 [2024-10-29 10:59:25.668112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:20.405 [2024-10-29 10:59:25.668168] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:20.405 [2024-10-29 10:59:25.668222] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:20.405 [2024-10-29 10:59:25.668243] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:20.405 [2024-10-29 10:59:25.668261] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:20.405 [2024-10-29 10:59:25.668279] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:20.405 [2024-10-29 10:59:25.668291] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:10:20.405 request: 00:10:20.405 { 00:10:20.405 "name": "raid_bdev1", 00:10:20.405 "raid_level": "raid0", 00:10:20.405 "base_bdevs": [ 00:10:20.406 "malloc1", 00:10:20.406 "malloc2", 00:10:20.406 "malloc3", 00:10:20.406 "malloc4" 00:10:20.406 ], 00:10:20.406 "strip_size_kb": 64, 00:10:20.406 "superblock": false, 00:10:20.406 "method": "bdev_raid_create", 00:10:20.406 "req_id": 1 00:10:20.406 } 00:10:20.406 Got JSON-RPC error response 00:10:20.406 response: 00:10:20.406 { 00:10:20.406 "code": -17, 00:10:20.406 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:20.406 } 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.406 [2024-10-29 10:59:25.721354] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:20.406 [2024-10-29 10:59:25.721514] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:20.406 [2024-10-29 10:59:25.721561] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:20.406 [2024-10-29 10:59:25.721595] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:20.406 [2024-10-29 10:59:25.724290] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:20.406 [2024-10-29 10:59:25.724378] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:20.406 [2024-10-29 10:59:25.724527] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:20.406 [2024-10-29 10:59:25.724612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:20.406 pt1 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.406 "name": "raid_bdev1", 00:10:20.406 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:20.406 "strip_size_kb": 64, 00:10:20.406 "state": "configuring", 00:10:20.406 "raid_level": "raid0", 00:10:20.406 "superblock": true, 00:10:20.406 "num_base_bdevs": 4, 00:10:20.406 "num_base_bdevs_discovered": 1, 00:10:20.406 "num_base_bdevs_operational": 4, 00:10:20.406 "base_bdevs_list": [ 00:10:20.406 { 00:10:20.406 "name": "pt1", 00:10:20.406 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:20.406 "is_configured": true, 00:10:20.406 "data_offset": 2048, 00:10:20.406 "data_size": 63488 00:10:20.406 }, 00:10:20.406 { 00:10:20.406 "name": null, 00:10:20.406 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:20.406 "is_configured": false, 00:10:20.406 "data_offset": 2048, 00:10:20.406 "data_size": 63488 00:10:20.406 }, 00:10:20.406 { 00:10:20.406 "name": null, 00:10:20.406 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:20.406 "is_configured": false, 00:10:20.406 "data_offset": 2048, 00:10:20.406 "data_size": 63488 00:10:20.406 }, 00:10:20.406 { 00:10:20.406 "name": null, 00:10:20.406 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:20.406 "is_configured": false, 00:10:20.406 "data_offset": 2048, 00:10:20.406 "data_size": 63488 00:10:20.406 } 00:10:20.406 ] 00:10:20.406 }' 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.406 10:59:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.977 [2024-10-29 10:59:26.204589] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:20.977 [2024-10-29 10:59:26.204686] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:20.977 [2024-10-29 10:59:26.204715] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:20.977 [2024-10-29 10:59:26.204727] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:20.977 [2024-10-29 10:59:26.205258] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:20.977 [2024-10-29 10:59:26.205293] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:20.977 [2024-10-29 10:59:26.205417] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:20.977 [2024-10-29 10:59:26.205474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:20.977 pt2 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.977 [2024-10-29 10:59:26.216567] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.977 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.978 "name": "raid_bdev1", 00:10:20.978 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:20.978 "strip_size_kb": 64, 00:10:20.978 "state": "configuring", 00:10:20.978 "raid_level": "raid0", 00:10:20.978 "superblock": true, 00:10:20.978 "num_base_bdevs": 4, 00:10:20.978 "num_base_bdevs_discovered": 1, 00:10:20.978 "num_base_bdevs_operational": 4, 00:10:20.978 "base_bdevs_list": [ 00:10:20.978 { 00:10:20.978 "name": "pt1", 00:10:20.978 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:20.978 "is_configured": true, 00:10:20.978 "data_offset": 2048, 00:10:20.978 "data_size": 63488 00:10:20.978 }, 00:10:20.978 { 00:10:20.978 "name": null, 00:10:20.978 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:20.978 "is_configured": false, 00:10:20.978 "data_offset": 0, 00:10:20.978 "data_size": 63488 00:10:20.978 }, 00:10:20.978 { 00:10:20.978 "name": null, 00:10:20.978 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:20.978 "is_configured": false, 00:10:20.978 "data_offset": 2048, 00:10:20.978 "data_size": 63488 00:10:20.978 }, 00:10:20.978 { 00:10:20.978 "name": null, 00:10:20.978 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:20.978 "is_configured": false, 00:10:20.978 "data_offset": 2048, 00:10:20.978 "data_size": 63488 00:10:20.978 } 00:10:20.978 ] 00:10:20.978 }' 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.978 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.239 [2024-10-29 10:59:26.671872] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:21.239 [2024-10-29 10:59:26.672065] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.239 [2024-10-29 10:59:26.672107] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:21.239 [2024-10-29 10:59:26.672177] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.239 [2024-10-29 10:59:26.672747] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.239 [2024-10-29 10:59:26.672819] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:21.239 [2024-10-29 10:59:26.672963] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:21.239 [2024-10-29 10:59:26.673024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:21.239 pt2 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.239 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.240 [2024-10-29 10:59:26.683792] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:21.240 [2024-10-29 10:59:26.683899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.240 [2024-10-29 10:59:26.683940] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:21.240 [2024-10-29 10:59:26.683975] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.240 [2024-10-29 10:59:26.684475] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.240 [2024-10-29 10:59:26.684547] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:21.240 [2024-10-29 10:59:26.684679] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:21.240 [2024-10-29 10:59:26.684738] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:21.240 pt3 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.240 [2024-10-29 10:59:26.695741] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:21.240 [2024-10-29 10:59:26.695800] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:21.240 [2024-10-29 10:59:26.695819] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:21.240 [2024-10-29 10:59:26.695831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:21.240 [2024-10-29 10:59:26.696226] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:21.240 [2024-10-29 10:59:26.696260] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:21.240 [2024-10-29 10:59:26.696345] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:21.240 [2024-10-29 10:59:26.696386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:21.240 [2024-10-29 10:59:26.696508] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:21.240 [2024-10-29 10:59:26.696530] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:21.240 [2024-10-29 10:59:26.696818] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:21.240 [2024-10-29 10:59:26.696964] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:21.240 [2024-10-29 10:59:26.696973] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:10:21.240 [2024-10-29 10:59:26.697087] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:21.240 pt4 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.240 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.500 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.500 "name": "raid_bdev1", 00:10:21.500 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:21.500 "strip_size_kb": 64, 00:10:21.500 "state": "online", 00:10:21.500 "raid_level": "raid0", 00:10:21.500 "superblock": true, 00:10:21.500 "num_base_bdevs": 4, 00:10:21.500 "num_base_bdevs_discovered": 4, 00:10:21.500 "num_base_bdevs_operational": 4, 00:10:21.500 "base_bdevs_list": [ 00:10:21.500 { 00:10:21.500 "name": "pt1", 00:10:21.500 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:21.500 "is_configured": true, 00:10:21.500 "data_offset": 2048, 00:10:21.500 "data_size": 63488 00:10:21.500 }, 00:10:21.500 { 00:10:21.500 "name": "pt2", 00:10:21.500 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:21.500 "is_configured": true, 00:10:21.500 "data_offset": 2048, 00:10:21.500 "data_size": 63488 00:10:21.500 }, 00:10:21.500 { 00:10:21.501 "name": "pt3", 00:10:21.501 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:21.501 "is_configured": true, 00:10:21.501 "data_offset": 2048, 00:10:21.501 "data_size": 63488 00:10:21.501 }, 00:10:21.501 { 00:10:21.501 "name": "pt4", 00:10:21.501 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:21.501 "is_configured": true, 00:10:21.501 "data_offset": 2048, 00:10:21.501 "data_size": 63488 00:10:21.501 } 00:10:21.501 ] 00:10:21.501 }' 00:10:21.501 10:59:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.501 10:59:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.761 [2024-10-29 10:59:27.171952] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:21.761 "name": "raid_bdev1", 00:10:21.761 "aliases": [ 00:10:21.761 "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730" 00:10:21.761 ], 00:10:21.761 "product_name": "Raid Volume", 00:10:21.761 "block_size": 512, 00:10:21.761 "num_blocks": 253952, 00:10:21.761 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:21.761 "assigned_rate_limits": { 00:10:21.761 "rw_ios_per_sec": 0, 00:10:21.761 "rw_mbytes_per_sec": 0, 00:10:21.761 "r_mbytes_per_sec": 0, 00:10:21.761 "w_mbytes_per_sec": 0 00:10:21.761 }, 00:10:21.761 "claimed": false, 00:10:21.761 "zoned": false, 00:10:21.761 "supported_io_types": { 00:10:21.761 "read": true, 00:10:21.761 "write": true, 00:10:21.761 "unmap": true, 00:10:21.761 "flush": true, 00:10:21.761 "reset": true, 00:10:21.761 "nvme_admin": false, 00:10:21.761 "nvme_io": false, 00:10:21.761 "nvme_io_md": false, 00:10:21.761 "write_zeroes": true, 00:10:21.761 "zcopy": false, 00:10:21.761 "get_zone_info": false, 00:10:21.761 "zone_management": false, 00:10:21.761 "zone_append": false, 00:10:21.761 "compare": false, 00:10:21.761 "compare_and_write": false, 00:10:21.761 "abort": false, 00:10:21.761 "seek_hole": false, 00:10:21.761 "seek_data": false, 00:10:21.761 "copy": false, 00:10:21.761 "nvme_iov_md": false 00:10:21.761 }, 00:10:21.761 "memory_domains": [ 00:10:21.761 { 00:10:21.761 "dma_device_id": "system", 00:10:21.761 "dma_device_type": 1 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.761 "dma_device_type": 2 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "system", 00:10:21.761 "dma_device_type": 1 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.761 "dma_device_type": 2 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "system", 00:10:21.761 "dma_device_type": 1 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.761 "dma_device_type": 2 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "system", 00:10:21.761 "dma_device_type": 1 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.761 "dma_device_type": 2 00:10:21.761 } 00:10:21.761 ], 00:10:21.761 "driver_specific": { 00:10:21.761 "raid": { 00:10:21.761 "uuid": "bd5cd3c1-861a-45d4-afbe-5c1fcdde3730", 00:10:21.761 "strip_size_kb": 64, 00:10:21.761 "state": "online", 00:10:21.761 "raid_level": "raid0", 00:10:21.761 "superblock": true, 00:10:21.761 "num_base_bdevs": 4, 00:10:21.761 "num_base_bdevs_discovered": 4, 00:10:21.761 "num_base_bdevs_operational": 4, 00:10:21.761 "base_bdevs_list": [ 00:10:21.761 { 00:10:21.761 "name": "pt1", 00:10:21.761 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:21.761 "is_configured": true, 00:10:21.761 "data_offset": 2048, 00:10:21.761 "data_size": 63488 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "name": "pt2", 00:10:21.761 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:21.761 "is_configured": true, 00:10:21.761 "data_offset": 2048, 00:10:21.761 "data_size": 63488 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "name": "pt3", 00:10:21.761 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:21.761 "is_configured": true, 00:10:21.761 "data_offset": 2048, 00:10:21.761 "data_size": 63488 00:10:21.761 }, 00:10:21.761 { 00:10:21.761 "name": "pt4", 00:10:21.761 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:21.761 "is_configured": true, 00:10:21.761 "data_offset": 2048, 00:10:21.761 "data_size": 63488 00:10:21.761 } 00:10:21.761 ] 00:10:21.761 } 00:10:21.761 } 00:10:21.761 }' 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:21.761 pt2 00:10:21.761 pt3 00:10:21.761 pt4' 00:10:21.761 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.021 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.021 [2024-10-29 10:59:27.499886] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' bd5cd3c1-861a-45d4-afbe-5c1fcdde3730 '!=' bd5cd3c1-861a-45d4-afbe-5c1fcdde3730 ']' 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81807 00:10:22.280 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 81807 ']' 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 81807 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 81807 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 81807' 00:10:22.281 killing process with pid 81807 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 81807 00:10:22.281 [2024-10-29 10:59:27.580915] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:22.281 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 81807 00:10:22.281 [2024-10-29 10:59:27.581105] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:22.281 [2024-10-29 10:59:27.581202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:22.281 [2024-10-29 10:59:27.581277] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:10:22.281 [2024-10-29 10:59:27.664046] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:22.540 10:59:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:22.540 00:10:22.540 real 0m4.431s 00:10:22.540 user 0m6.833s 00:10:22.540 sys 0m1.030s 00:10:22.540 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:22.540 10:59:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.540 ************************************ 00:10:22.540 END TEST raid_superblock_test 00:10:22.540 ************************************ 00:10:22.800 10:59:28 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:10:22.800 10:59:28 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:22.800 10:59:28 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:22.800 10:59:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:22.800 ************************************ 00:10:22.800 START TEST raid_read_error_test 00:10:22.800 ************************************ 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 4 read 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.J4muIXwweB 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82061 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82061 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 82061 ']' 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:22.800 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:22.800 10:59:28 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.800 [2024-10-29 10:59:28.168651] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:22.800 [2024-10-29 10:59:28.168792] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82061 ] 00:10:23.059 [2024-10-29 10:59:28.317620] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:23.059 [2024-10-29 10:59:28.357145] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:23.059 [2024-10-29 10:59:28.433494] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:23.059 [2024-10-29 10:59:28.433534] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 BaseBdev1_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 true 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 [2024-10-29 10:59:29.048086] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:23.651 [2024-10-29 10:59:29.048224] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:23.651 [2024-10-29 10:59:29.048250] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:23.651 [2024-10-29 10:59:29.048267] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:23.651 [2024-10-29 10:59:29.050722] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:23.651 [2024-10-29 10:59:29.050761] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:23.651 BaseBdev1 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 BaseBdev2_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 true 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.651 [2024-10-29 10:59:29.094909] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:23.651 [2024-10-29 10:59:29.095039] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:23.651 [2024-10-29 10:59:29.095064] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:23.651 [2024-10-29 10:59:29.095073] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:23.651 [2024-10-29 10:59:29.097510] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:23.651 [2024-10-29 10:59:29.097546] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:23.651 BaseBdev2 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:23.651 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.652 BaseBdev3_malloc 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.652 true 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.652 [2024-10-29 10:59:29.141481] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:23.652 [2024-10-29 10:59:29.141531] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:23.652 [2024-10-29 10:59:29.141551] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:23.652 [2024-10-29 10:59:29.141561] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:23.652 [2024-10-29 10:59:29.143944] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:23.652 [2024-10-29 10:59:29.143980] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:23.652 BaseBdev3 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.652 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.912 BaseBdev4_malloc 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.912 true 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.912 [2024-10-29 10:59:29.196761] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:23.912 [2024-10-29 10:59:29.196889] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:23.912 [2024-10-29 10:59:29.196918] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:23.912 [2024-10-29 10:59:29.196928] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:23.912 [2024-10-29 10:59:29.199281] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:23.912 [2024-10-29 10:59:29.199319] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:23.912 BaseBdev4 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.912 [2024-10-29 10:59:29.208804] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:23.912 [2024-10-29 10:59:29.210928] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:23.912 [2024-10-29 10:59:29.211012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:23.912 [2024-10-29 10:59:29.211064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:23.912 [2024-10-29 10:59:29.211264] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:23.912 [2024-10-29 10:59:29.211276] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:23.912 [2024-10-29 10:59:29.211596] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:23.912 [2024-10-29 10:59:29.211737] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:23.912 [2024-10-29 10:59:29.211757] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:23.912 [2024-10-29 10:59:29.211912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.912 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.912 "name": "raid_bdev1", 00:10:23.912 "uuid": "3331c2d9-e773-4000-ac2d-297862b60524", 00:10:23.912 "strip_size_kb": 64, 00:10:23.912 "state": "online", 00:10:23.912 "raid_level": "raid0", 00:10:23.912 "superblock": true, 00:10:23.912 "num_base_bdevs": 4, 00:10:23.912 "num_base_bdevs_discovered": 4, 00:10:23.912 "num_base_bdevs_operational": 4, 00:10:23.912 "base_bdevs_list": [ 00:10:23.912 { 00:10:23.912 "name": "BaseBdev1", 00:10:23.912 "uuid": "13bb3475-16dd-5dc2-bb51-1ca44434d778", 00:10:23.912 "is_configured": true, 00:10:23.912 "data_offset": 2048, 00:10:23.912 "data_size": 63488 00:10:23.912 }, 00:10:23.912 { 00:10:23.912 "name": "BaseBdev2", 00:10:23.912 "uuid": "f4e77cb6-87b8-5d11-b819-bcae026581da", 00:10:23.913 "is_configured": true, 00:10:23.913 "data_offset": 2048, 00:10:23.913 "data_size": 63488 00:10:23.913 }, 00:10:23.913 { 00:10:23.913 "name": "BaseBdev3", 00:10:23.913 "uuid": "3283f6f0-68d4-5f03-b25d-ba1d8d4bcaf5", 00:10:23.913 "is_configured": true, 00:10:23.913 "data_offset": 2048, 00:10:23.913 "data_size": 63488 00:10:23.913 }, 00:10:23.913 { 00:10:23.913 "name": "BaseBdev4", 00:10:23.913 "uuid": "3c03b0eb-aa24-5422-b256-de7c86203bb9", 00:10:23.913 "is_configured": true, 00:10:23.913 "data_offset": 2048, 00:10:23.913 "data_size": 63488 00:10:23.913 } 00:10:23.913 ] 00:10:23.913 }' 00:10:23.913 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.913 10:59:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.173 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:24.173 10:59:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:24.432 [2024-10-29 10:59:29.752528] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.370 "name": "raid_bdev1", 00:10:25.370 "uuid": "3331c2d9-e773-4000-ac2d-297862b60524", 00:10:25.370 "strip_size_kb": 64, 00:10:25.370 "state": "online", 00:10:25.370 "raid_level": "raid0", 00:10:25.370 "superblock": true, 00:10:25.370 "num_base_bdevs": 4, 00:10:25.370 "num_base_bdevs_discovered": 4, 00:10:25.370 "num_base_bdevs_operational": 4, 00:10:25.370 "base_bdevs_list": [ 00:10:25.370 { 00:10:25.370 "name": "BaseBdev1", 00:10:25.370 "uuid": "13bb3475-16dd-5dc2-bb51-1ca44434d778", 00:10:25.370 "is_configured": true, 00:10:25.370 "data_offset": 2048, 00:10:25.370 "data_size": 63488 00:10:25.370 }, 00:10:25.370 { 00:10:25.370 "name": "BaseBdev2", 00:10:25.370 "uuid": "f4e77cb6-87b8-5d11-b819-bcae026581da", 00:10:25.370 "is_configured": true, 00:10:25.370 "data_offset": 2048, 00:10:25.370 "data_size": 63488 00:10:25.370 }, 00:10:25.370 { 00:10:25.370 "name": "BaseBdev3", 00:10:25.370 "uuid": "3283f6f0-68d4-5f03-b25d-ba1d8d4bcaf5", 00:10:25.370 "is_configured": true, 00:10:25.370 "data_offset": 2048, 00:10:25.370 "data_size": 63488 00:10:25.370 }, 00:10:25.370 { 00:10:25.370 "name": "BaseBdev4", 00:10:25.370 "uuid": "3c03b0eb-aa24-5422-b256-de7c86203bb9", 00:10:25.370 "is_configured": true, 00:10:25.370 "data_offset": 2048, 00:10:25.370 "data_size": 63488 00:10:25.370 } 00:10:25.370 ] 00:10:25.370 }' 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.370 10:59:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.630 [2024-10-29 10:59:31.081118] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:25.630 [2024-10-29 10:59:31.081259] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:25.630 [2024-10-29 10:59:31.083824] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:25.630 [2024-10-29 10:59:31.083940] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:25.630 [2024-10-29 10:59:31.084017] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:25.630 [2024-10-29 10:59:31.084074] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82061 00:10:25.630 { 00:10:25.630 "results": [ 00:10:25.630 { 00:10:25.630 "job": "raid_bdev1", 00:10:25.630 "core_mask": "0x1", 00:10:25.630 "workload": "randrw", 00:10:25.630 "percentage": 50, 00:10:25.630 "status": "finished", 00:10:25.630 "queue_depth": 1, 00:10:25.630 "io_size": 131072, 00:10:25.630 "runtime": 1.329087, 00:10:25.630 "iops": 14250.383910157876, 00:10:25.630 "mibps": 1781.2979887697345, 00:10:25.630 "io_failed": 1, 00:10:25.630 "io_timeout": 0, 00:10:25.630 "avg_latency_us": 98.8096904914341, 00:10:25.630 "min_latency_us": 25.9353711790393, 00:10:25.630 "max_latency_us": 1373.6803493449781 00:10:25.630 } 00:10:25.630 ], 00:10:25.630 "core_count": 1 00:10:25.630 } 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 82061 ']' 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 82061 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 82061 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 82061' 00:10:25.630 killing process with pid 82061 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 82061 00:10:25.630 [2024-10-29 10:59:31.123434] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:25.630 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 82061 00:10:25.889 [2024-10-29 10:59:31.191526] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.J4muIXwweB 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:10:26.148 00:10:26.148 real 0m3.463s 00:10:26.148 user 0m4.199s 00:10:26.148 sys 0m0.633s 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:26.148 10:59:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.148 ************************************ 00:10:26.148 END TEST raid_read_error_test 00:10:26.148 ************************************ 00:10:26.148 10:59:31 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:10:26.148 10:59:31 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:26.148 10:59:31 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:26.148 10:59:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:26.148 ************************************ 00:10:26.148 START TEST raid_write_error_test 00:10:26.148 ************************************ 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid0 4 write 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:26.148 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.pElsFI9lQV 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82195 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82195 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 82195 ']' 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:26.149 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:26.149 10:59:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.408 [2024-10-29 10:59:31.695919] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:26.408 [2024-10-29 10:59:31.696168] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82195 ] 00:10:26.408 [2024-10-29 10:59:31.868020] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:26.408 [2024-10-29 10:59:31.907421] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:26.667 [2024-10-29 10:59:31.985166] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:26.667 [2024-10-29 10:59:31.985229] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.232 BaseBdev1_malloc 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.232 true 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.232 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.232 [2024-10-29 10:59:32.576209] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:27.232 [2024-10-29 10:59:32.576372] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.232 [2024-10-29 10:59:32.576425] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:27.232 [2024-10-29 10:59:32.576460] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.232 [2024-10-29 10:59:32.579063] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.233 [2024-10-29 10:59:32.579139] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:27.233 BaseBdev1 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 BaseBdev2_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 true 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 [2024-10-29 10:59:32.619975] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:27.233 [2024-10-29 10:59:32.620106] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.233 [2024-10-29 10:59:32.620145] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:27.233 [2024-10-29 10:59:32.620176] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.233 [2024-10-29 10:59:32.622788] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.233 [2024-10-29 10:59:32.622865] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:27.233 BaseBdev2 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 BaseBdev3_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 true 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 [2024-10-29 10:59:32.666910] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:27.233 [2024-10-29 10:59:32.666966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.233 [2024-10-29 10:59:32.666987] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:27.233 [2024-10-29 10:59:32.666997] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.233 [2024-10-29 10:59:32.669354] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.233 [2024-10-29 10:59:32.669411] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:27.233 BaseBdev3 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 BaseBdev4_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 true 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.233 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.233 [2024-10-29 10:59:32.725615] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:27.233 [2024-10-29 10:59:32.725737] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:27.233 [2024-10-29 10:59:32.725764] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:27.233 [2024-10-29 10:59:32.725774] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:27.233 [2024-10-29 10:59:32.728129] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:27.233 [2024-10-29 10:59:32.728167] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:27.493 BaseBdev4 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.493 [2024-10-29 10:59:32.737677] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:27.493 [2024-10-29 10:59:32.739925] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:27.493 [2024-10-29 10:59:32.740061] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:27.493 [2024-10-29 10:59:32.740142] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:27.493 [2024-10-29 10:59:32.740425] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:27.493 [2024-10-29 10:59:32.740477] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:27.493 [2024-10-29 10:59:32.740799] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:27.493 [2024-10-29 10:59:32.740990] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:27.493 [2024-10-29 10:59:32.741037] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:27.493 [2024-10-29 10:59:32.741254] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.493 "name": "raid_bdev1", 00:10:27.493 "uuid": "ca7f2bfc-23c4-47f7-9040-8dbffeca8267", 00:10:27.493 "strip_size_kb": 64, 00:10:27.493 "state": "online", 00:10:27.493 "raid_level": "raid0", 00:10:27.493 "superblock": true, 00:10:27.493 "num_base_bdevs": 4, 00:10:27.493 "num_base_bdevs_discovered": 4, 00:10:27.493 "num_base_bdevs_operational": 4, 00:10:27.493 "base_bdevs_list": [ 00:10:27.493 { 00:10:27.493 "name": "BaseBdev1", 00:10:27.493 "uuid": "4d35c1d5-b04c-5230-8857-9cfd13375d6a", 00:10:27.493 "is_configured": true, 00:10:27.493 "data_offset": 2048, 00:10:27.493 "data_size": 63488 00:10:27.493 }, 00:10:27.493 { 00:10:27.493 "name": "BaseBdev2", 00:10:27.493 "uuid": "58bde5b1-b0ea-5b93-9431-74fe3678018a", 00:10:27.493 "is_configured": true, 00:10:27.493 "data_offset": 2048, 00:10:27.493 "data_size": 63488 00:10:27.493 }, 00:10:27.493 { 00:10:27.493 "name": "BaseBdev3", 00:10:27.493 "uuid": "a2f4c948-de3f-52ab-b913-560a39adb3b4", 00:10:27.493 "is_configured": true, 00:10:27.493 "data_offset": 2048, 00:10:27.493 "data_size": 63488 00:10:27.493 }, 00:10:27.493 { 00:10:27.493 "name": "BaseBdev4", 00:10:27.493 "uuid": "aa3a5cb5-bc54-5966-b162-581fb02d806c", 00:10:27.493 "is_configured": true, 00:10:27.493 "data_offset": 2048, 00:10:27.493 "data_size": 63488 00:10:27.493 } 00:10:27.493 ] 00:10:27.493 }' 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.493 10:59:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.752 10:59:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:27.752 10:59:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:27.752 [2024-10-29 10:59:33.221350] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:28.691 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.692 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.951 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.952 "name": "raid_bdev1", 00:10:28.952 "uuid": "ca7f2bfc-23c4-47f7-9040-8dbffeca8267", 00:10:28.952 "strip_size_kb": 64, 00:10:28.952 "state": "online", 00:10:28.952 "raid_level": "raid0", 00:10:28.952 "superblock": true, 00:10:28.952 "num_base_bdevs": 4, 00:10:28.952 "num_base_bdevs_discovered": 4, 00:10:28.952 "num_base_bdevs_operational": 4, 00:10:28.952 "base_bdevs_list": [ 00:10:28.952 { 00:10:28.952 "name": "BaseBdev1", 00:10:28.952 "uuid": "4d35c1d5-b04c-5230-8857-9cfd13375d6a", 00:10:28.952 "is_configured": true, 00:10:28.952 "data_offset": 2048, 00:10:28.952 "data_size": 63488 00:10:28.952 }, 00:10:28.952 { 00:10:28.952 "name": "BaseBdev2", 00:10:28.952 "uuid": "58bde5b1-b0ea-5b93-9431-74fe3678018a", 00:10:28.952 "is_configured": true, 00:10:28.952 "data_offset": 2048, 00:10:28.952 "data_size": 63488 00:10:28.952 }, 00:10:28.952 { 00:10:28.952 "name": "BaseBdev3", 00:10:28.952 "uuid": "a2f4c948-de3f-52ab-b913-560a39adb3b4", 00:10:28.952 "is_configured": true, 00:10:28.952 "data_offset": 2048, 00:10:28.952 "data_size": 63488 00:10:28.952 }, 00:10:28.952 { 00:10:28.952 "name": "BaseBdev4", 00:10:28.952 "uuid": "aa3a5cb5-bc54-5966-b162-581fb02d806c", 00:10:28.952 "is_configured": true, 00:10:28.952 "data_offset": 2048, 00:10:28.952 "data_size": 63488 00:10:28.952 } 00:10:28.952 ] 00:10:28.952 }' 00:10:28.952 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.952 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.212 [2024-10-29 10:59:34.598418] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:29.212 [2024-10-29 10:59:34.598532] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:29.212 [2024-10-29 10:59:34.601056] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:29.212 { 00:10:29.212 "results": [ 00:10:29.212 { 00:10:29.212 "job": "raid_bdev1", 00:10:29.212 "core_mask": "0x1", 00:10:29.212 "workload": "randrw", 00:10:29.212 "percentage": 50, 00:10:29.212 "status": "finished", 00:10:29.212 "queue_depth": 1, 00:10:29.212 "io_size": 131072, 00:10:29.212 "runtime": 1.377524, 00:10:29.212 "iops": 14193.582108188315, 00:10:29.212 "mibps": 1774.1977635235394, 00:10:29.212 "io_failed": 1, 00:10:29.212 "io_timeout": 0, 00:10:29.212 "avg_latency_us": 99.24233161375072, 00:10:29.212 "min_latency_us": 25.3764192139738, 00:10:29.212 "max_latency_us": 1359.3711790393013 00:10:29.212 } 00:10:29.212 ], 00:10:29.212 "core_count": 1 00:10:29.212 } 00:10:29.212 [2024-10-29 10:59:34.601154] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:29.212 [2024-10-29 10:59:34.601232] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:29.212 [2024-10-29 10:59:34.601243] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82195 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 82195 ']' 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 82195 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 82195 00:10:29.212 killing process with pid 82195 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 82195' 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 82195 00:10:29.212 [2024-10-29 10:59:34.645733] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:29.212 10:59:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 82195 00:10:29.472 [2024-10-29 10:59:34.712323] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.pElsFI9lQV 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:29.733 00:10:29.733 real 0m3.448s 00:10:29.733 user 0m4.195s 00:10:29.733 sys 0m0.630s 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:29.733 10:59:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.733 ************************************ 00:10:29.733 END TEST raid_write_error_test 00:10:29.733 ************************************ 00:10:29.733 10:59:35 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:29.733 10:59:35 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:10:29.733 10:59:35 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:29.733 10:59:35 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:29.733 10:59:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:29.733 ************************************ 00:10:29.733 START TEST raid_state_function_test 00:10:29.733 ************************************ 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 4 false 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:29.733 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:29.734 Process raid pid: 82328 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=82328 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82328' 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 82328 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 82328 ']' 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:29.734 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:29.734 10:59:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.734 [2024-10-29 10:59:35.213486] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:29.734 [2024-10-29 10:59:35.213727] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:29.994 [2024-10-29 10:59:35.385268] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:29.994 [2024-10-29 10:59:35.423979] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:30.254 [2024-10-29 10:59:35.500692] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:30.254 [2024-10-29 10:59:35.500839] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.825 [2024-10-29 10:59:36.053239] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:30.825 [2024-10-29 10:59:36.053384] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:30.825 [2024-10-29 10:59:36.053439] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:30.825 [2024-10-29 10:59:36.053465] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:30.825 [2024-10-29 10:59:36.053504] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:30.825 [2024-10-29 10:59:36.053529] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:30.825 [2024-10-29 10:59:36.053587] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:30.825 [2024-10-29 10:59:36.053627] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.825 "name": "Existed_Raid", 00:10:30.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.825 "strip_size_kb": 64, 00:10:30.825 "state": "configuring", 00:10:30.825 "raid_level": "concat", 00:10:30.825 "superblock": false, 00:10:30.825 "num_base_bdevs": 4, 00:10:30.825 "num_base_bdevs_discovered": 0, 00:10:30.825 "num_base_bdevs_operational": 4, 00:10:30.825 "base_bdevs_list": [ 00:10:30.825 { 00:10:30.825 "name": "BaseBdev1", 00:10:30.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.825 "is_configured": false, 00:10:30.825 "data_offset": 0, 00:10:30.825 "data_size": 0 00:10:30.825 }, 00:10:30.825 { 00:10:30.825 "name": "BaseBdev2", 00:10:30.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.825 "is_configured": false, 00:10:30.825 "data_offset": 0, 00:10:30.825 "data_size": 0 00:10:30.825 }, 00:10:30.825 { 00:10:30.825 "name": "BaseBdev3", 00:10:30.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.825 "is_configured": false, 00:10:30.825 "data_offset": 0, 00:10:30.825 "data_size": 0 00:10:30.825 }, 00:10:30.825 { 00:10:30.825 "name": "BaseBdev4", 00:10:30.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.825 "is_configured": false, 00:10:30.825 "data_offset": 0, 00:10:30.825 "data_size": 0 00:10:30.825 } 00:10:30.825 ] 00:10:30.825 }' 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.825 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 [2024-10-29 10:59:36.516376] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:31.086 [2024-10-29 10:59:36.516509] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 [2024-10-29 10:59:36.528318] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:31.086 [2024-10-29 10:59:36.528421] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:31.086 [2024-10-29 10:59:36.528450] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:31.086 [2024-10-29 10:59:36.528473] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:31.086 [2024-10-29 10:59:36.528490] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:31.086 [2024-10-29 10:59:36.528511] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:31.086 [2024-10-29 10:59:36.528527] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:31.086 [2024-10-29 10:59:36.528548] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 [2024-10-29 10:59:36.555222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:31.086 BaseBdev1 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.086 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.086 [ 00:10:31.086 { 00:10:31.086 "name": "BaseBdev1", 00:10:31.086 "aliases": [ 00:10:31.086 "a2993f7a-886b-4b49-b042-b11405d3a133" 00:10:31.086 ], 00:10:31.086 "product_name": "Malloc disk", 00:10:31.086 "block_size": 512, 00:10:31.086 "num_blocks": 65536, 00:10:31.086 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:31.086 "assigned_rate_limits": { 00:10:31.086 "rw_ios_per_sec": 0, 00:10:31.086 "rw_mbytes_per_sec": 0, 00:10:31.086 "r_mbytes_per_sec": 0, 00:10:31.086 "w_mbytes_per_sec": 0 00:10:31.086 }, 00:10:31.086 "claimed": true, 00:10:31.086 "claim_type": "exclusive_write", 00:10:31.086 "zoned": false, 00:10:31.086 "supported_io_types": { 00:10:31.086 "read": true, 00:10:31.347 "write": true, 00:10:31.347 "unmap": true, 00:10:31.347 "flush": true, 00:10:31.347 "reset": true, 00:10:31.347 "nvme_admin": false, 00:10:31.347 "nvme_io": false, 00:10:31.347 "nvme_io_md": false, 00:10:31.347 "write_zeroes": true, 00:10:31.347 "zcopy": true, 00:10:31.347 "get_zone_info": false, 00:10:31.347 "zone_management": false, 00:10:31.347 "zone_append": false, 00:10:31.347 "compare": false, 00:10:31.347 "compare_and_write": false, 00:10:31.347 "abort": true, 00:10:31.347 "seek_hole": false, 00:10:31.347 "seek_data": false, 00:10:31.347 "copy": true, 00:10:31.347 "nvme_iov_md": false 00:10:31.347 }, 00:10:31.347 "memory_domains": [ 00:10:31.347 { 00:10:31.347 "dma_device_id": "system", 00:10:31.347 "dma_device_type": 1 00:10:31.347 }, 00:10:31.347 { 00:10:31.347 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.347 "dma_device_type": 2 00:10:31.347 } 00:10:31.347 ], 00:10:31.347 "driver_specific": {} 00:10:31.347 } 00:10:31.347 ] 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.347 "name": "Existed_Raid", 00:10:31.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.347 "strip_size_kb": 64, 00:10:31.347 "state": "configuring", 00:10:31.347 "raid_level": "concat", 00:10:31.347 "superblock": false, 00:10:31.347 "num_base_bdevs": 4, 00:10:31.347 "num_base_bdevs_discovered": 1, 00:10:31.347 "num_base_bdevs_operational": 4, 00:10:31.347 "base_bdevs_list": [ 00:10:31.347 { 00:10:31.347 "name": "BaseBdev1", 00:10:31.347 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:31.347 "is_configured": true, 00:10:31.347 "data_offset": 0, 00:10:31.347 "data_size": 65536 00:10:31.347 }, 00:10:31.347 { 00:10:31.347 "name": "BaseBdev2", 00:10:31.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.347 "is_configured": false, 00:10:31.347 "data_offset": 0, 00:10:31.347 "data_size": 0 00:10:31.347 }, 00:10:31.347 { 00:10:31.347 "name": "BaseBdev3", 00:10:31.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.347 "is_configured": false, 00:10:31.347 "data_offset": 0, 00:10:31.347 "data_size": 0 00:10:31.347 }, 00:10:31.347 { 00:10:31.347 "name": "BaseBdev4", 00:10:31.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.347 "is_configured": false, 00:10:31.347 "data_offset": 0, 00:10:31.347 "data_size": 0 00:10:31.347 } 00:10:31.347 ] 00:10:31.347 }' 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.347 10:59:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.607 [2024-10-29 10:59:37.062472] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:31.607 [2024-10-29 10:59:37.062632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.607 [2024-10-29 10:59:37.074455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:31.607 [2024-10-29 10:59:37.076794] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:31.607 [2024-10-29 10:59:37.076879] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:31.607 [2024-10-29 10:59:37.076907] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:31.607 [2024-10-29 10:59:37.076929] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:31.607 [2024-10-29 10:59:37.076946] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:31.607 [2024-10-29 10:59:37.076967] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.607 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.608 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.868 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.868 "name": "Existed_Raid", 00:10:31.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.868 "strip_size_kb": 64, 00:10:31.868 "state": "configuring", 00:10:31.868 "raid_level": "concat", 00:10:31.868 "superblock": false, 00:10:31.868 "num_base_bdevs": 4, 00:10:31.868 "num_base_bdevs_discovered": 1, 00:10:31.868 "num_base_bdevs_operational": 4, 00:10:31.868 "base_bdevs_list": [ 00:10:31.868 { 00:10:31.868 "name": "BaseBdev1", 00:10:31.868 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:31.868 "is_configured": true, 00:10:31.868 "data_offset": 0, 00:10:31.868 "data_size": 65536 00:10:31.868 }, 00:10:31.868 { 00:10:31.868 "name": "BaseBdev2", 00:10:31.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.868 "is_configured": false, 00:10:31.868 "data_offset": 0, 00:10:31.868 "data_size": 0 00:10:31.868 }, 00:10:31.868 { 00:10:31.868 "name": "BaseBdev3", 00:10:31.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.868 "is_configured": false, 00:10:31.868 "data_offset": 0, 00:10:31.868 "data_size": 0 00:10:31.868 }, 00:10:31.868 { 00:10:31.868 "name": "BaseBdev4", 00:10:31.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.868 "is_configured": false, 00:10:31.868 "data_offset": 0, 00:10:31.868 "data_size": 0 00:10:31.868 } 00:10:31.868 ] 00:10:31.868 }' 00:10:31.868 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.868 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.128 [2024-10-29 10:59:37.526562] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:32.128 BaseBdev2 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:32.128 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.129 [ 00:10:32.129 { 00:10:32.129 "name": "BaseBdev2", 00:10:32.129 "aliases": [ 00:10:32.129 "1593d063-f61f-4b8c-92a6-8a75446afeb4" 00:10:32.129 ], 00:10:32.129 "product_name": "Malloc disk", 00:10:32.129 "block_size": 512, 00:10:32.129 "num_blocks": 65536, 00:10:32.129 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:32.129 "assigned_rate_limits": { 00:10:32.129 "rw_ios_per_sec": 0, 00:10:32.129 "rw_mbytes_per_sec": 0, 00:10:32.129 "r_mbytes_per_sec": 0, 00:10:32.129 "w_mbytes_per_sec": 0 00:10:32.129 }, 00:10:32.129 "claimed": true, 00:10:32.129 "claim_type": "exclusive_write", 00:10:32.129 "zoned": false, 00:10:32.129 "supported_io_types": { 00:10:32.129 "read": true, 00:10:32.129 "write": true, 00:10:32.129 "unmap": true, 00:10:32.129 "flush": true, 00:10:32.129 "reset": true, 00:10:32.129 "nvme_admin": false, 00:10:32.129 "nvme_io": false, 00:10:32.129 "nvme_io_md": false, 00:10:32.129 "write_zeroes": true, 00:10:32.129 "zcopy": true, 00:10:32.129 "get_zone_info": false, 00:10:32.129 "zone_management": false, 00:10:32.129 "zone_append": false, 00:10:32.129 "compare": false, 00:10:32.129 "compare_and_write": false, 00:10:32.129 "abort": true, 00:10:32.129 "seek_hole": false, 00:10:32.129 "seek_data": false, 00:10:32.129 "copy": true, 00:10:32.129 "nvme_iov_md": false 00:10:32.129 }, 00:10:32.129 "memory_domains": [ 00:10:32.129 { 00:10:32.129 "dma_device_id": "system", 00:10:32.129 "dma_device_type": 1 00:10:32.129 }, 00:10:32.129 { 00:10:32.129 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.129 "dma_device_type": 2 00:10:32.129 } 00:10:32.129 ], 00:10:32.129 "driver_specific": {} 00:10:32.129 } 00:10:32.129 ] 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.129 "name": "Existed_Raid", 00:10:32.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.129 "strip_size_kb": 64, 00:10:32.129 "state": "configuring", 00:10:32.129 "raid_level": "concat", 00:10:32.129 "superblock": false, 00:10:32.129 "num_base_bdevs": 4, 00:10:32.129 "num_base_bdevs_discovered": 2, 00:10:32.129 "num_base_bdevs_operational": 4, 00:10:32.129 "base_bdevs_list": [ 00:10:32.129 { 00:10:32.129 "name": "BaseBdev1", 00:10:32.129 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:32.129 "is_configured": true, 00:10:32.129 "data_offset": 0, 00:10:32.129 "data_size": 65536 00:10:32.129 }, 00:10:32.129 { 00:10:32.129 "name": "BaseBdev2", 00:10:32.129 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:32.129 "is_configured": true, 00:10:32.129 "data_offset": 0, 00:10:32.129 "data_size": 65536 00:10:32.129 }, 00:10:32.129 { 00:10:32.129 "name": "BaseBdev3", 00:10:32.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.129 "is_configured": false, 00:10:32.129 "data_offset": 0, 00:10:32.129 "data_size": 0 00:10:32.129 }, 00:10:32.129 { 00:10:32.129 "name": "BaseBdev4", 00:10:32.129 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.129 "is_configured": false, 00:10:32.129 "data_offset": 0, 00:10:32.129 "data_size": 0 00:10:32.129 } 00:10:32.129 ] 00:10:32.129 }' 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.129 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.700 10:59:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:32.700 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.700 10:59:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.700 [2024-10-29 10:59:38.009859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:32.700 BaseBdev3 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.700 [ 00:10:32.700 { 00:10:32.700 "name": "BaseBdev3", 00:10:32.700 "aliases": [ 00:10:32.700 "00e6f9db-4762-43dc-b259-21b7a4b04c85" 00:10:32.700 ], 00:10:32.700 "product_name": "Malloc disk", 00:10:32.700 "block_size": 512, 00:10:32.700 "num_blocks": 65536, 00:10:32.700 "uuid": "00e6f9db-4762-43dc-b259-21b7a4b04c85", 00:10:32.700 "assigned_rate_limits": { 00:10:32.700 "rw_ios_per_sec": 0, 00:10:32.700 "rw_mbytes_per_sec": 0, 00:10:32.700 "r_mbytes_per_sec": 0, 00:10:32.700 "w_mbytes_per_sec": 0 00:10:32.700 }, 00:10:32.700 "claimed": true, 00:10:32.700 "claim_type": "exclusive_write", 00:10:32.700 "zoned": false, 00:10:32.700 "supported_io_types": { 00:10:32.700 "read": true, 00:10:32.700 "write": true, 00:10:32.700 "unmap": true, 00:10:32.700 "flush": true, 00:10:32.700 "reset": true, 00:10:32.700 "nvme_admin": false, 00:10:32.700 "nvme_io": false, 00:10:32.700 "nvme_io_md": false, 00:10:32.700 "write_zeroes": true, 00:10:32.700 "zcopy": true, 00:10:32.700 "get_zone_info": false, 00:10:32.700 "zone_management": false, 00:10:32.700 "zone_append": false, 00:10:32.700 "compare": false, 00:10:32.700 "compare_and_write": false, 00:10:32.700 "abort": true, 00:10:32.700 "seek_hole": false, 00:10:32.700 "seek_data": false, 00:10:32.700 "copy": true, 00:10:32.700 "nvme_iov_md": false 00:10:32.700 }, 00:10:32.700 "memory_domains": [ 00:10:32.700 { 00:10:32.700 "dma_device_id": "system", 00:10:32.700 "dma_device_type": 1 00:10:32.700 }, 00:10:32.700 { 00:10:32.700 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.700 "dma_device_type": 2 00:10:32.700 } 00:10:32.700 ], 00:10:32.700 "driver_specific": {} 00:10:32.700 } 00:10:32.700 ] 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.700 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.701 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.701 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.701 "name": "Existed_Raid", 00:10:32.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.701 "strip_size_kb": 64, 00:10:32.701 "state": "configuring", 00:10:32.701 "raid_level": "concat", 00:10:32.701 "superblock": false, 00:10:32.701 "num_base_bdevs": 4, 00:10:32.701 "num_base_bdevs_discovered": 3, 00:10:32.701 "num_base_bdevs_operational": 4, 00:10:32.701 "base_bdevs_list": [ 00:10:32.701 { 00:10:32.701 "name": "BaseBdev1", 00:10:32.701 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:32.701 "is_configured": true, 00:10:32.701 "data_offset": 0, 00:10:32.701 "data_size": 65536 00:10:32.701 }, 00:10:32.701 { 00:10:32.701 "name": "BaseBdev2", 00:10:32.701 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:32.701 "is_configured": true, 00:10:32.701 "data_offset": 0, 00:10:32.701 "data_size": 65536 00:10:32.701 }, 00:10:32.701 { 00:10:32.701 "name": "BaseBdev3", 00:10:32.701 "uuid": "00e6f9db-4762-43dc-b259-21b7a4b04c85", 00:10:32.701 "is_configured": true, 00:10:32.701 "data_offset": 0, 00:10:32.701 "data_size": 65536 00:10:32.701 }, 00:10:32.701 { 00:10:32.701 "name": "BaseBdev4", 00:10:32.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.701 "is_configured": false, 00:10:32.701 "data_offset": 0, 00:10:32.701 "data_size": 0 00:10:32.701 } 00:10:32.701 ] 00:10:32.701 }' 00:10:32.701 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.701 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.961 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:32.961 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.961 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.221 [2024-10-29 10:59:38.470407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:33.222 [2024-10-29 10:59:38.470549] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:33.222 [2024-10-29 10:59:38.470575] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:33.222 [2024-10-29 10:59:38.470961] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:33.222 [2024-10-29 10:59:38.471174] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:33.222 [2024-10-29 10:59:38.471222] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:33.222 [2024-10-29 10:59:38.471557] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:33.222 BaseBdev4 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.222 [ 00:10:33.222 { 00:10:33.222 "name": "BaseBdev4", 00:10:33.222 "aliases": [ 00:10:33.222 "f065948e-20ca-4821-b22c-acaa1a71e907" 00:10:33.222 ], 00:10:33.222 "product_name": "Malloc disk", 00:10:33.222 "block_size": 512, 00:10:33.222 "num_blocks": 65536, 00:10:33.222 "uuid": "f065948e-20ca-4821-b22c-acaa1a71e907", 00:10:33.222 "assigned_rate_limits": { 00:10:33.222 "rw_ios_per_sec": 0, 00:10:33.222 "rw_mbytes_per_sec": 0, 00:10:33.222 "r_mbytes_per_sec": 0, 00:10:33.222 "w_mbytes_per_sec": 0 00:10:33.222 }, 00:10:33.222 "claimed": true, 00:10:33.222 "claim_type": "exclusive_write", 00:10:33.222 "zoned": false, 00:10:33.222 "supported_io_types": { 00:10:33.222 "read": true, 00:10:33.222 "write": true, 00:10:33.222 "unmap": true, 00:10:33.222 "flush": true, 00:10:33.222 "reset": true, 00:10:33.222 "nvme_admin": false, 00:10:33.222 "nvme_io": false, 00:10:33.222 "nvme_io_md": false, 00:10:33.222 "write_zeroes": true, 00:10:33.222 "zcopy": true, 00:10:33.222 "get_zone_info": false, 00:10:33.222 "zone_management": false, 00:10:33.222 "zone_append": false, 00:10:33.222 "compare": false, 00:10:33.222 "compare_and_write": false, 00:10:33.222 "abort": true, 00:10:33.222 "seek_hole": false, 00:10:33.222 "seek_data": false, 00:10:33.222 "copy": true, 00:10:33.222 "nvme_iov_md": false 00:10:33.222 }, 00:10:33.222 "memory_domains": [ 00:10:33.222 { 00:10:33.222 "dma_device_id": "system", 00:10:33.222 "dma_device_type": 1 00:10:33.222 }, 00:10:33.222 { 00:10:33.222 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.222 "dma_device_type": 2 00:10:33.222 } 00:10:33.222 ], 00:10:33.222 "driver_specific": {} 00:10:33.222 } 00:10:33.222 ] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.222 "name": "Existed_Raid", 00:10:33.222 "uuid": "4e8fdf7c-4d75-43a5-ab6a-95fcf81503b1", 00:10:33.222 "strip_size_kb": 64, 00:10:33.222 "state": "online", 00:10:33.222 "raid_level": "concat", 00:10:33.222 "superblock": false, 00:10:33.222 "num_base_bdevs": 4, 00:10:33.222 "num_base_bdevs_discovered": 4, 00:10:33.222 "num_base_bdevs_operational": 4, 00:10:33.222 "base_bdevs_list": [ 00:10:33.222 { 00:10:33.222 "name": "BaseBdev1", 00:10:33.222 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:33.222 "is_configured": true, 00:10:33.222 "data_offset": 0, 00:10:33.222 "data_size": 65536 00:10:33.222 }, 00:10:33.222 { 00:10:33.222 "name": "BaseBdev2", 00:10:33.222 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:33.222 "is_configured": true, 00:10:33.222 "data_offset": 0, 00:10:33.222 "data_size": 65536 00:10:33.222 }, 00:10:33.222 { 00:10:33.222 "name": "BaseBdev3", 00:10:33.222 "uuid": "00e6f9db-4762-43dc-b259-21b7a4b04c85", 00:10:33.222 "is_configured": true, 00:10:33.222 "data_offset": 0, 00:10:33.222 "data_size": 65536 00:10:33.222 }, 00:10:33.222 { 00:10:33.222 "name": "BaseBdev4", 00:10:33.222 "uuid": "f065948e-20ca-4821-b22c-acaa1a71e907", 00:10:33.222 "is_configured": true, 00:10:33.222 "data_offset": 0, 00:10:33.222 "data_size": 65536 00:10:33.222 } 00:10:33.222 ] 00:10:33.222 }' 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.222 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:33.483 [2024-10-29 10:59:38.898118] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.483 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:33.483 "name": "Existed_Raid", 00:10:33.483 "aliases": [ 00:10:33.483 "4e8fdf7c-4d75-43a5-ab6a-95fcf81503b1" 00:10:33.483 ], 00:10:33.483 "product_name": "Raid Volume", 00:10:33.483 "block_size": 512, 00:10:33.483 "num_blocks": 262144, 00:10:33.483 "uuid": "4e8fdf7c-4d75-43a5-ab6a-95fcf81503b1", 00:10:33.483 "assigned_rate_limits": { 00:10:33.483 "rw_ios_per_sec": 0, 00:10:33.483 "rw_mbytes_per_sec": 0, 00:10:33.483 "r_mbytes_per_sec": 0, 00:10:33.483 "w_mbytes_per_sec": 0 00:10:33.483 }, 00:10:33.483 "claimed": false, 00:10:33.483 "zoned": false, 00:10:33.483 "supported_io_types": { 00:10:33.483 "read": true, 00:10:33.483 "write": true, 00:10:33.483 "unmap": true, 00:10:33.483 "flush": true, 00:10:33.483 "reset": true, 00:10:33.483 "nvme_admin": false, 00:10:33.483 "nvme_io": false, 00:10:33.483 "nvme_io_md": false, 00:10:33.483 "write_zeroes": true, 00:10:33.483 "zcopy": false, 00:10:33.483 "get_zone_info": false, 00:10:33.483 "zone_management": false, 00:10:33.483 "zone_append": false, 00:10:33.483 "compare": false, 00:10:33.483 "compare_and_write": false, 00:10:33.483 "abort": false, 00:10:33.483 "seek_hole": false, 00:10:33.483 "seek_data": false, 00:10:33.483 "copy": false, 00:10:33.483 "nvme_iov_md": false 00:10:33.483 }, 00:10:33.483 "memory_domains": [ 00:10:33.483 { 00:10:33.483 "dma_device_id": "system", 00:10:33.483 "dma_device_type": 1 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.483 "dma_device_type": 2 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "system", 00:10:33.483 "dma_device_type": 1 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.483 "dma_device_type": 2 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "system", 00:10:33.483 "dma_device_type": 1 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.483 "dma_device_type": 2 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "system", 00:10:33.483 "dma_device_type": 1 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.483 "dma_device_type": 2 00:10:33.483 } 00:10:33.483 ], 00:10:33.483 "driver_specific": { 00:10:33.483 "raid": { 00:10:33.483 "uuid": "4e8fdf7c-4d75-43a5-ab6a-95fcf81503b1", 00:10:33.483 "strip_size_kb": 64, 00:10:33.483 "state": "online", 00:10:33.483 "raid_level": "concat", 00:10:33.483 "superblock": false, 00:10:33.483 "num_base_bdevs": 4, 00:10:33.483 "num_base_bdevs_discovered": 4, 00:10:33.483 "num_base_bdevs_operational": 4, 00:10:33.483 "base_bdevs_list": [ 00:10:33.483 { 00:10:33.483 "name": "BaseBdev1", 00:10:33.483 "uuid": "a2993f7a-886b-4b49-b042-b11405d3a133", 00:10:33.483 "is_configured": true, 00:10:33.483 "data_offset": 0, 00:10:33.483 "data_size": 65536 00:10:33.483 }, 00:10:33.483 { 00:10:33.483 "name": "BaseBdev2", 00:10:33.483 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:33.483 "is_configured": true, 00:10:33.483 "data_offset": 0, 00:10:33.483 "data_size": 65536 00:10:33.483 }, 00:10:33.484 { 00:10:33.484 "name": "BaseBdev3", 00:10:33.484 "uuid": "00e6f9db-4762-43dc-b259-21b7a4b04c85", 00:10:33.484 "is_configured": true, 00:10:33.484 "data_offset": 0, 00:10:33.484 "data_size": 65536 00:10:33.484 }, 00:10:33.484 { 00:10:33.484 "name": "BaseBdev4", 00:10:33.484 "uuid": "f065948e-20ca-4821-b22c-acaa1a71e907", 00:10:33.484 "is_configured": true, 00:10:33.484 "data_offset": 0, 00:10:33.484 "data_size": 65536 00:10:33.484 } 00:10:33.484 ] 00:10:33.484 } 00:10:33.484 } 00:10:33.484 }' 00:10:33.484 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:33.484 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:33.484 BaseBdev2 00:10:33.484 BaseBdev3 00:10:33.484 BaseBdev4' 00:10:33.484 10:59:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.744 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:33.744 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.745 [2024-10-29 10:59:39.209301] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:33.745 [2024-10-29 10:59:39.209348] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:33.745 [2024-10-29 10:59:39.209424] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.745 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.005 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.005 "name": "Existed_Raid", 00:10:34.005 "uuid": "4e8fdf7c-4d75-43a5-ab6a-95fcf81503b1", 00:10:34.005 "strip_size_kb": 64, 00:10:34.005 "state": "offline", 00:10:34.005 "raid_level": "concat", 00:10:34.005 "superblock": false, 00:10:34.005 "num_base_bdevs": 4, 00:10:34.005 "num_base_bdevs_discovered": 3, 00:10:34.005 "num_base_bdevs_operational": 3, 00:10:34.005 "base_bdevs_list": [ 00:10:34.005 { 00:10:34.005 "name": null, 00:10:34.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.005 "is_configured": false, 00:10:34.005 "data_offset": 0, 00:10:34.005 "data_size": 65536 00:10:34.005 }, 00:10:34.005 { 00:10:34.005 "name": "BaseBdev2", 00:10:34.005 "uuid": "1593d063-f61f-4b8c-92a6-8a75446afeb4", 00:10:34.005 "is_configured": true, 00:10:34.005 "data_offset": 0, 00:10:34.005 "data_size": 65536 00:10:34.005 }, 00:10:34.005 { 00:10:34.005 "name": "BaseBdev3", 00:10:34.005 "uuid": "00e6f9db-4762-43dc-b259-21b7a4b04c85", 00:10:34.005 "is_configured": true, 00:10:34.005 "data_offset": 0, 00:10:34.005 "data_size": 65536 00:10:34.005 }, 00:10:34.005 { 00:10:34.005 "name": "BaseBdev4", 00:10:34.005 "uuid": "f065948e-20ca-4821-b22c-acaa1a71e907", 00:10:34.005 "is_configured": true, 00:10:34.005 "data_offset": 0, 00:10:34.005 "data_size": 65536 00:10:34.005 } 00:10:34.005 ] 00:10:34.005 }' 00:10:34.005 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.005 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.266 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:34.266 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.266 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.266 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.267 [2024-10-29 10:59:39.717137] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.267 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 [2024-10-29 10:59:39.792884] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 [2024-10-29 10:59:39.865878] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:34.527 [2024-10-29 10:59:39.866026] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 BaseBdev2 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.527 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.527 [ 00:10:34.527 { 00:10:34.527 "name": "BaseBdev2", 00:10:34.527 "aliases": [ 00:10:34.527 "e7c4240d-be92-4d56-929c-5b66aae14300" 00:10:34.527 ], 00:10:34.527 "product_name": "Malloc disk", 00:10:34.527 "block_size": 512, 00:10:34.527 "num_blocks": 65536, 00:10:34.527 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:34.527 "assigned_rate_limits": { 00:10:34.527 "rw_ios_per_sec": 0, 00:10:34.527 "rw_mbytes_per_sec": 0, 00:10:34.527 "r_mbytes_per_sec": 0, 00:10:34.527 "w_mbytes_per_sec": 0 00:10:34.527 }, 00:10:34.527 "claimed": false, 00:10:34.527 "zoned": false, 00:10:34.527 "supported_io_types": { 00:10:34.527 "read": true, 00:10:34.527 "write": true, 00:10:34.527 "unmap": true, 00:10:34.527 "flush": true, 00:10:34.527 "reset": true, 00:10:34.527 "nvme_admin": false, 00:10:34.527 "nvme_io": false, 00:10:34.527 "nvme_io_md": false, 00:10:34.527 "write_zeroes": true, 00:10:34.527 "zcopy": true, 00:10:34.527 "get_zone_info": false, 00:10:34.527 "zone_management": false, 00:10:34.527 "zone_append": false, 00:10:34.527 "compare": false, 00:10:34.527 "compare_and_write": false, 00:10:34.527 "abort": true, 00:10:34.527 "seek_hole": false, 00:10:34.527 "seek_data": false, 00:10:34.527 "copy": true, 00:10:34.527 "nvme_iov_md": false 00:10:34.527 }, 00:10:34.527 "memory_domains": [ 00:10:34.527 { 00:10:34.527 "dma_device_id": "system", 00:10:34.528 "dma_device_type": 1 00:10:34.528 }, 00:10:34.528 { 00:10:34.528 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.528 "dma_device_type": 2 00:10:34.528 } 00:10:34.528 ], 00:10:34.528 "driver_specific": {} 00:10:34.528 } 00:10:34.528 ] 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.528 10:59:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.528 BaseBdev3 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:34.528 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.787 [ 00:10:34.787 { 00:10:34.787 "name": "BaseBdev3", 00:10:34.787 "aliases": [ 00:10:34.787 "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79" 00:10:34.787 ], 00:10:34.787 "product_name": "Malloc disk", 00:10:34.787 "block_size": 512, 00:10:34.787 "num_blocks": 65536, 00:10:34.787 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:34.787 "assigned_rate_limits": { 00:10:34.787 "rw_ios_per_sec": 0, 00:10:34.787 "rw_mbytes_per_sec": 0, 00:10:34.787 "r_mbytes_per_sec": 0, 00:10:34.787 "w_mbytes_per_sec": 0 00:10:34.787 }, 00:10:34.787 "claimed": false, 00:10:34.787 "zoned": false, 00:10:34.787 "supported_io_types": { 00:10:34.787 "read": true, 00:10:34.787 "write": true, 00:10:34.787 "unmap": true, 00:10:34.787 "flush": true, 00:10:34.787 "reset": true, 00:10:34.787 "nvme_admin": false, 00:10:34.787 "nvme_io": false, 00:10:34.787 "nvme_io_md": false, 00:10:34.787 "write_zeroes": true, 00:10:34.787 "zcopy": true, 00:10:34.787 "get_zone_info": false, 00:10:34.787 "zone_management": false, 00:10:34.787 "zone_append": false, 00:10:34.787 "compare": false, 00:10:34.787 "compare_and_write": false, 00:10:34.787 "abort": true, 00:10:34.787 "seek_hole": false, 00:10:34.787 "seek_data": false, 00:10:34.787 "copy": true, 00:10:34.787 "nvme_iov_md": false 00:10:34.787 }, 00:10:34.787 "memory_domains": [ 00:10:34.787 { 00:10:34.787 "dma_device_id": "system", 00:10:34.787 "dma_device_type": 1 00:10:34.787 }, 00:10:34.787 { 00:10:34.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.787 "dma_device_type": 2 00:10:34.787 } 00:10:34.787 ], 00:10:34.787 "driver_specific": {} 00:10:34.787 } 00:10:34.787 ] 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.787 BaseBdev4 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:34.787 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.788 [ 00:10:34.788 { 00:10:34.788 "name": "BaseBdev4", 00:10:34.788 "aliases": [ 00:10:34.788 "44b322df-c28c-48df-9239-87996deefbd5" 00:10:34.788 ], 00:10:34.788 "product_name": "Malloc disk", 00:10:34.788 "block_size": 512, 00:10:34.788 "num_blocks": 65536, 00:10:34.788 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:34.788 "assigned_rate_limits": { 00:10:34.788 "rw_ios_per_sec": 0, 00:10:34.788 "rw_mbytes_per_sec": 0, 00:10:34.788 "r_mbytes_per_sec": 0, 00:10:34.788 "w_mbytes_per_sec": 0 00:10:34.788 }, 00:10:34.788 "claimed": false, 00:10:34.788 "zoned": false, 00:10:34.788 "supported_io_types": { 00:10:34.788 "read": true, 00:10:34.788 "write": true, 00:10:34.788 "unmap": true, 00:10:34.788 "flush": true, 00:10:34.788 "reset": true, 00:10:34.788 "nvme_admin": false, 00:10:34.788 "nvme_io": false, 00:10:34.788 "nvme_io_md": false, 00:10:34.788 "write_zeroes": true, 00:10:34.788 "zcopy": true, 00:10:34.788 "get_zone_info": false, 00:10:34.788 "zone_management": false, 00:10:34.788 "zone_append": false, 00:10:34.788 "compare": false, 00:10:34.788 "compare_and_write": false, 00:10:34.788 "abort": true, 00:10:34.788 "seek_hole": false, 00:10:34.788 "seek_data": false, 00:10:34.788 "copy": true, 00:10:34.788 "nvme_iov_md": false 00:10:34.788 }, 00:10:34.788 "memory_domains": [ 00:10:34.788 { 00:10:34.788 "dma_device_id": "system", 00:10:34.788 "dma_device_type": 1 00:10:34.788 }, 00:10:34.788 { 00:10:34.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.788 "dma_device_type": 2 00:10:34.788 } 00:10:34.788 ], 00:10:34.788 "driver_specific": {} 00:10:34.788 } 00:10:34.788 ] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.788 [2024-10-29 10:59:40.116578] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:34.788 [2024-10-29 10:59:40.116716] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:34.788 [2024-10-29 10:59:40.116774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:34.788 [2024-10-29 10:59:40.119113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:34.788 [2024-10-29 10:59:40.119222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.788 "name": "Existed_Raid", 00:10:34.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.788 "strip_size_kb": 64, 00:10:34.788 "state": "configuring", 00:10:34.788 "raid_level": "concat", 00:10:34.788 "superblock": false, 00:10:34.788 "num_base_bdevs": 4, 00:10:34.788 "num_base_bdevs_discovered": 3, 00:10:34.788 "num_base_bdevs_operational": 4, 00:10:34.788 "base_bdevs_list": [ 00:10:34.788 { 00:10:34.788 "name": "BaseBdev1", 00:10:34.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.788 "is_configured": false, 00:10:34.788 "data_offset": 0, 00:10:34.788 "data_size": 0 00:10:34.788 }, 00:10:34.788 { 00:10:34.788 "name": "BaseBdev2", 00:10:34.788 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:34.788 "is_configured": true, 00:10:34.788 "data_offset": 0, 00:10:34.788 "data_size": 65536 00:10:34.788 }, 00:10:34.788 { 00:10:34.788 "name": "BaseBdev3", 00:10:34.788 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:34.788 "is_configured": true, 00:10:34.788 "data_offset": 0, 00:10:34.788 "data_size": 65536 00:10:34.788 }, 00:10:34.788 { 00:10:34.788 "name": "BaseBdev4", 00:10:34.788 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:34.788 "is_configured": true, 00:10:34.788 "data_offset": 0, 00:10:34.788 "data_size": 65536 00:10:34.788 } 00:10:34.788 ] 00:10:34.788 }' 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.788 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.047 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:35.047 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.047 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.047 [2024-10-29 10:59:40.543853] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.306 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.306 "name": "Existed_Raid", 00:10:35.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.306 "strip_size_kb": 64, 00:10:35.306 "state": "configuring", 00:10:35.306 "raid_level": "concat", 00:10:35.306 "superblock": false, 00:10:35.306 "num_base_bdevs": 4, 00:10:35.306 "num_base_bdevs_discovered": 2, 00:10:35.306 "num_base_bdevs_operational": 4, 00:10:35.306 "base_bdevs_list": [ 00:10:35.306 { 00:10:35.306 "name": "BaseBdev1", 00:10:35.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.306 "is_configured": false, 00:10:35.307 "data_offset": 0, 00:10:35.307 "data_size": 0 00:10:35.307 }, 00:10:35.307 { 00:10:35.307 "name": null, 00:10:35.307 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:35.307 "is_configured": false, 00:10:35.307 "data_offset": 0, 00:10:35.307 "data_size": 65536 00:10:35.307 }, 00:10:35.307 { 00:10:35.307 "name": "BaseBdev3", 00:10:35.307 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:35.307 "is_configured": true, 00:10:35.307 "data_offset": 0, 00:10:35.307 "data_size": 65536 00:10:35.307 }, 00:10:35.307 { 00:10:35.307 "name": "BaseBdev4", 00:10:35.307 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:35.307 "is_configured": true, 00:10:35.307 "data_offset": 0, 00:10:35.307 "data_size": 65536 00:10:35.307 } 00:10:35.307 ] 00:10:35.307 }' 00:10:35.307 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.307 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.566 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:35.566 10:59:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.566 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.566 10:59:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.566 [2024-10-29 10:59:41.044294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:35.566 BaseBdev1 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:35.566 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.567 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.826 [ 00:10:35.826 { 00:10:35.826 "name": "BaseBdev1", 00:10:35.826 "aliases": [ 00:10:35.826 "2b5d6684-19bc-494b-9ec9-a29461616cb4" 00:10:35.826 ], 00:10:35.826 "product_name": "Malloc disk", 00:10:35.826 "block_size": 512, 00:10:35.826 "num_blocks": 65536, 00:10:35.826 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:35.826 "assigned_rate_limits": { 00:10:35.826 "rw_ios_per_sec": 0, 00:10:35.826 "rw_mbytes_per_sec": 0, 00:10:35.826 "r_mbytes_per_sec": 0, 00:10:35.826 "w_mbytes_per_sec": 0 00:10:35.826 }, 00:10:35.826 "claimed": true, 00:10:35.826 "claim_type": "exclusive_write", 00:10:35.826 "zoned": false, 00:10:35.826 "supported_io_types": { 00:10:35.826 "read": true, 00:10:35.826 "write": true, 00:10:35.826 "unmap": true, 00:10:35.826 "flush": true, 00:10:35.826 "reset": true, 00:10:35.826 "nvme_admin": false, 00:10:35.826 "nvme_io": false, 00:10:35.826 "nvme_io_md": false, 00:10:35.826 "write_zeroes": true, 00:10:35.826 "zcopy": true, 00:10:35.826 "get_zone_info": false, 00:10:35.826 "zone_management": false, 00:10:35.826 "zone_append": false, 00:10:35.826 "compare": false, 00:10:35.826 "compare_and_write": false, 00:10:35.826 "abort": true, 00:10:35.826 "seek_hole": false, 00:10:35.826 "seek_data": false, 00:10:35.826 "copy": true, 00:10:35.826 "nvme_iov_md": false 00:10:35.826 }, 00:10:35.826 "memory_domains": [ 00:10:35.826 { 00:10:35.826 "dma_device_id": "system", 00:10:35.826 "dma_device_type": 1 00:10:35.826 }, 00:10:35.826 { 00:10:35.826 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.826 "dma_device_type": 2 00:10:35.826 } 00:10:35.826 ], 00:10:35.826 "driver_specific": {} 00:10:35.826 } 00:10:35.826 ] 00:10:35.826 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.826 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:35.826 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.827 "name": "Existed_Raid", 00:10:35.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.827 "strip_size_kb": 64, 00:10:35.827 "state": "configuring", 00:10:35.827 "raid_level": "concat", 00:10:35.827 "superblock": false, 00:10:35.827 "num_base_bdevs": 4, 00:10:35.827 "num_base_bdevs_discovered": 3, 00:10:35.827 "num_base_bdevs_operational": 4, 00:10:35.827 "base_bdevs_list": [ 00:10:35.827 { 00:10:35.827 "name": "BaseBdev1", 00:10:35.827 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:35.827 "is_configured": true, 00:10:35.827 "data_offset": 0, 00:10:35.827 "data_size": 65536 00:10:35.827 }, 00:10:35.827 { 00:10:35.827 "name": null, 00:10:35.827 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:35.827 "is_configured": false, 00:10:35.827 "data_offset": 0, 00:10:35.827 "data_size": 65536 00:10:35.827 }, 00:10:35.827 { 00:10:35.827 "name": "BaseBdev3", 00:10:35.827 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:35.827 "is_configured": true, 00:10:35.827 "data_offset": 0, 00:10:35.827 "data_size": 65536 00:10:35.827 }, 00:10:35.827 { 00:10:35.827 "name": "BaseBdev4", 00:10:35.827 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:35.827 "is_configured": true, 00:10:35.827 "data_offset": 0, 00:10:35.827 "data_size": 65536 00:10:35.827 } 00:10:35.827 ] 00:10:35.827 }' 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.827 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.087 [2024-10-29 10:59:41.571621] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.087 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.347 "name": "Existed_Raid", 00:10:36.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.347 "strip_size_kb": 64, 00:10:36.347 "state": "configuring", 00:10:36.347 "raid_level": "concat", 00:10:36.347 "superblock": false, 00:10:36.347 "num_base_bdevs": 4, 00:10:36.347 "num_base_bdevs_discovered": 2, 00:10:36.347 "num_base_bdevs_operational": 4, 00:10:36.347 "base_bdevs_list": [ 00:10:36.347 { 00:10:36.347 "name": "BaseBdev1", 00:10:36.347 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:36.347 "is_configured": true, 00:10:36.347 "data_offset": 0, 00:10:36.347 "data_size": 65536 00:10:36.347 }, 00:10:36.347 { 00:10:36.347 "name": null, 00:10:36.347 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:36.347 "is_configured": false, 00:10:36.347 "data_offset": 0, 00:10:36.347 "data_size": 65536 00:10:36.347 }, 00:10:36.347 { 00:10:36.347 "name": null, 00:10:36.347 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:36.347 "is_configured": false, 00:10:36.347 "data_offset": 0, 00:10:36.347 "data_size": 65536 00:10:36.347 }, 00:10:36.347 { 00:10:36.347 "name": "BaseBdev4", 00:10:36.347 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:36.347 "is_configured": true, 00:10:36.347 "data_offset": 0, 00:10:36.347 "data_size": 65536 00:10:36.347 } 00:10:36.347 ] 00:10:36.347 }' 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.347 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.608 10:59:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.608 [2024-10-29 10:59:41.998996] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.608 "name": "Existed_Raid", 00:10:36.608 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.608 "strip_size_kb": 64, 00:10:36.608 "state": "configuring", 00:10:36.608 "raid_level": "concat", 00:10:36.608 "superblock": false, 00:10:36.608 "num_base_bdevs": 4, 00:10:36.608 "num_base_bdevs_discovered": 3, 00:10:36.608 "num_base_bdevs_operational": 4, 00:10:36.608 "base_bdevs_list": [ 00:10:36.608 { 00:10:36.608 "name": "BaseBdev1", 00:10:36.608 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:36.608 "is_configured": true, 00:10:36.608 "data_offset": 0, 00:10:36.608 "data_size": 65536 00:10:36.608 }, 00:10:36.608 { 00:10:36.608 "name": null, 00:10:36.608 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:36.608 "is_configured": false, 00:10:36.608 "data_offset": 0, 00:10:36.608 "data_size": 65536 00:10:36.608 }, 00:10:36.608 { 00:10:36.608 "name": "BaseBdev3", 00:10:36.608 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:36.608 "is_configured": true, 00:10:36.608 "data_offset": 0, 00:10:36.608 "data_size": 65536 00:10:36.608 }, 00:10:36.608 { 00:10:36.608 "name": "BaseBdev4", 00:10:36.608 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:36.608 "is_configured": true, 00:10:36.608 "data_offset": 0, 00:10:36.608 "data_size": 65536 00:10:36.608 } 00:10:36.608 ] 00:10:36.608 }' 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.608 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.180 [2024-10-29 10:59:42.466237] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.180 "name": "Existed_Raid", 00:10:37.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:37.180 "strip_size_kb": 64, 00:10:37.180 "state": "configuring", 00:10:37.180 "raid_level": "concat", 00:10:37.180 "superblock": false, 00:10:37.180 "num_base_bdevs": 4, 00:10:37.180 "num_base_bdevs_discovered": 2, 00:10:37.180 "num_base_bdevs_operational": 4, 00:10:37.180 "base_bdevs_list": [ 00:10:37.180 { 00:10:37.180 "name": null, 00:10:37.180 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:37.180 "is_configured": false, 00:10:37.180 "data_offset": 0, 00:10:37.180 "data_size": 65536 00:10:37.180 }, 00:10:37.180 { 00:10:37.180 "name": null, 00:10:37.180 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:37.180 "is_configured": false, 00:10:37.180 "data_offset": 0, 00:10:37.180 "data_size": 65536 00:10:37.180 }, 00:10:37.180 { 00:10:37.180 "name": "BaseBdev3", 00:10:37.180 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:37.180 "is_configured": true, 00:10:37.180 "data_offset": 0, 00:10:37.180 "data_size": 65536 00:10:37.180 }, 00:10:37.180 { 00:10:37.180 "name": "BaseBdev4", 00:10:37.180 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:37.180 "is_configured": true, 00:10:37.180 "data_offset": 0, 00:10:37.180 "data_size": 65536 00:10:37.180 } 00:10:37.180 ] 00:10:37.180 }' 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.180 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.440 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.440 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:37.440 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.440 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.700 [2024-10-29 10:59:42.973335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.700 10:59:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.700 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.700 "name": "Existed_Raid", 00:10:37.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:37.700 "strip_size_kb": 64, 00:10:37.700 "state": "configuring", 00:10:37.700 "raid_level": "concat", 00:10:37.700 "superblock": false, 00:10:37.700 "num_base_bdevs": 4, 00:10:37.700 "num_base_bdevs_discovered": 3, 00:10:37.700 "num_base_bdevs_operational": 4, 00:10:37.700 "base_bdevs_list": [ 00:10:37.700 { 00:10:37.700 "name": null, 00:10:37.700 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:37.700 "is_configured": false, 00:10:37.700 "data_offset": 0, 00:10:37.700 "data_size": 65536 00:10:37.700 }, 00:10:37.700 { 00:10:37.700 "name": "BaseBdev2", 00:10:37.700 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:37.700 "is_configured": true, 00:10:37.700 "data_offset": 0, 00:10:37.701 "data_size": 65536 00:10:37.701 }, 00:10:37.701 { 00:10:37.701 "name": "BaseBdev3", 00:10:37.701 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:37.701 "is_configured": true, 00:10:37.701 "data_offset": 0, 00:10:37.701 "data_size": 65536 00:10:37.701 }, 00:10:37.701 { 00:10:37.701 "name": "BaseBdev4", 00:10:37.701 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:37.701 "is_configured": true, 00:10:37.701 "data_offset": 0, 00:10:37.701 "data_size": 65536 00:10:37.701 } 00:10:37.701 ] 00:10:37.701 }' 00:10:37.701 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.701 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.960 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:37.961 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2b5d6684-19bc-494b-9ec9-a29461616cb4 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.220 [2024-10-29 10:59:43.489363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:38.220 [2024-10-29 10:59:43.489507] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:38.220 [2024-10-29 10:59:43.489531] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:38.220 [2024-10-29 10:59:43.489908] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:38.220 [2024-10-29 10:59:43.490100] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:38.220 [2024-10-29 10:59:43.490144] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:38.220 [2024-10-29 10:59:43.490439] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:38.220 NewBaseBdev 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.220 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.220 [ 00:10:38.220 { 00:10:38.220 "name": "NewBaseBdev", 00:10:38.220 "aliases": [ 00:10:38.220 "2b5d6684-19bc-494b-9ec9-a29461616cb4" 00:10:38.220 ], 00:10:38.220 "product_name": "Malloc disk", 00:10:38.220 "block_size": 512, 00:10:38.220 "num_blocks": 65536, 00:10:38.220 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:38.221 "assigned_rate_limits": { 00:10:38.221 "rw_ios_per_sec": 0, 00:10:38.221 "rw_mbytes_per_sec": 0, 00:10:38.221 "r_mbytes_per_sec": 0, 00:10:38.221 "w_mbytes_per_sec": 0 00:10:38.221 }, 00:10:38.221 "claimed": true, 00:10:38.221 "claim_type": "exclusive_write", 00:10:38.221 "zoned": false, 00:10:38.221 "supported_io_types": { 00:10:38.221 "read": true, 00:10:38.221 "write": true, 00:10:38.221 "unmap": true, 00:10:38.221 "flush": true, 00:10:38.221 "reset": true, 00:10:38.221 "nvme_admin": false, 00:10:38.221 "nvme_io": false, 00:10:38.221 "nvme_io_md": false, 00:10:38.221 "write_zeroes": true, 00:10:38.221 "zcopy": true, 00:10:38.221 "get_zone_info": false, 00:10:38.221 "zone_management": false, 00:10:38.221 "zone_append": false, 00:10:38.221 "compare": false, 00:10:38.221 "compare_and_write": false, 00:10:38.221 "abort": true, 00:10:38.221 "seek_hole": false, 00:10:38.221 "seek_data": false, 00:10:38.221 "copy": true, 00:10:38.221 "nvme_iov_md": false 00:10:38.221 }, 00:10:38.221 "memory_domains": [ 00:10:38.221 { 00:10:38.221 "dma_device_id": "system", 00:10:38.221 "dma_device_type": 1 00:10:38.221 }, 00:10:38.221 { 00:10:38.221 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.221 "dma_device_type": 2 00:10:38.221 } 00:10:38.221 ], 00:10:38.221 "driver_specific": {} 00:10:38.221 } 00:10:38.221 ] 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.221 "name": "Existed_Raid", 00:10:38.221 "uuid": "38db5c2e-1061-48ac-9f3a-2e2bf9e4821c", 00:10:38.221 "strip_size_kb": 64, 00:10:38.221 "state": "online", 00:10:38.221 "raid_level": "concat", 00:10:38.221 "superblock": false, 00:10:38.221 "num_base_bdevs": 4, 00:10:38.221 "num_base_bdevs_discovered": 4, 00:10:38.221 "num_base_bdevs_operational": 4, 00:10:38.221 "base_bdevs_list": [ 00:10:38.221 { 00:10:38.221 "name": "NewBaseBdev", 00:10:38.221 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:38.221 "is_configured": true, 00:10:38.221 "data_offset": 0, 00:10:38.221 "data_size": 65536 00:10:38.221 }, 00:10:38.221 { 00:10:38.221 "name": "BaseBdev2", 00:10:38.221 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:38.221 "is_configured": true, 00:10:38.221 "data_offset": 0, 00:10:38.221 "data_size": 65536 00:10:38.221 }, 00:10:38.221 { 00:10:38.221 "name": "BaseBdev3", 00:10:38.221 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:38.221 "is_configured": true, 00:10:38.221 "data_offset": 0, 00:10:38.221 "data_size": 65536 00:10:38.221 }, 00:10:38.221 { 00:10:38.221 "name": "BaseBdev4", 00:10:38.221 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:38.221 "is_configured": true, 00:10:38.221 "data_offset": 0, 00:10:38.221 "data_size": 65536 00:10:38.221 } 00:10:38.221 ] 00:10:38.221 }' 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.221 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.481 [2024-10-29 10:59:43.905089] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.481 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:38.481 "name": "Existed_Raid", 00:10:38.481 "aliases": [ 00:10:38.481 "38db5c2e-1061-48ac-9f3a-2e2bf9e4821c" 00:10:38.481 ], 00:10:38.481 "product_name": "Raid Volume", 00:10:38.481 "block_size": 512, 00:10:38.481 "num_blocks": 262144, 00:10:38.481 "uuid": "38db5c2e-1061-48ac-9f3a-2e2bf9e4821c", 00:10:38.481 "assigned_rate_limits": { 00:10:38.481 "rw_ios_per_sec": 0, 00:10:38.481 "rw_mbytes_per_sec": 0, 00:10:38.481 "r_mbytes_per_sec": 0, 00:10:38.481 "w_mbytes_per_sec": 0 00:10:38.481 }, 00:10:38.481 "claimed": false, 00:10:38.481 "zoned": false, 00:10:38.481 "supported_io_types": { 00:10:38.481 "read": true, 00:10:38.481 "write": true, 00:10:38.481 "unmap": true, 00:10:38.481 "flush": true, 00:10:38.481 "reset": true, 00:10:38.481 "nvme_admin": false, 00:10:38.481 "nvme_io": false, 00:10:38.481 "nvme_io_md": false, 00:10:38.481 "write_zeroes": true, 00:10:38.481 "zcopy": false, 00:10:38.481 "get_zone_info": false, 00:10:38.481 "zone_management": false, 00:10:38.481 "zone_append": false, 00:10:38.481 "compare": false, 00:10:38.481 "compare_and_write": false, 00:10:38.481 "abort": false, 00:10:38.481 "seek_hole": false, 00:10:38.481 "seek_data": false, 00:10:38.481 "copy": false, 00:10:38.481 "nvme_iov_md": false 00:10:38.481 }, 00:10:38.481 "memory_domains": [ 00:10:38.481 { 00:10:38.481 "dma_device_id": "system", 00:10:38.481 "dma_device_type": 1 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.481 "dma_device_type": 2 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "system", 00:10:38.481 "dma_device_type": 1 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.481 "dma_device_type": 2 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "system", 00:10:38.481 "dma_device_type": 1 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.481 "dma_device_type": 2 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "system", 00:10:38.481 "dma_device_type": 1 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.481 "dma_device_type": 2 00:10:38.481 } 00:10:38.481 ], 00:10:38.481 "driver_specific": { 00:10:38.481 "raid": { 00:10:38.481 "uuid": "38db5c2e-1061-48ac-9f3a-2e2bf9e4821c", 00:10:38.481 "strip_size_kb": 64, 00:10:38.481 "state": "online", 00:10:38.481 "raid_level": "concat", 00:10:38.481 "superblock": false, 00:10:38.481 "num_base_bdevs": 4, 00:10:38.481 "num_base_bdevs_discovered": 4, 00:10:38.481 "num_base_bdevs_operational": 4, 00:10:38.481 "base_bdevs_list": [ 00:10:38.481 { 00:10:38.481 "name": "NewBaseBdev", 00:10:38.481 "uuid": "2b5d6684-19bc-494b-9ec9-a29461616cb4", 00:10:38.481 "is_configured": true, 00:10:38.481 "data_offset": 0, 00:10:38.481 "data_size": 65536 00:10:38.481 }, 00:10:38.481 { 00:10:38.481 "name": "BaseBdev2", 00:10:38.481 "uuid": "e7c4240d-be92-4d56-929c-5b66aae14300", 00:10:38.481 "is_configured": true, 00:10:38.481 "data_offset": 0, 00:10:38.481 "data_size": 65536 00:10:38.482 }, 00:10:38.482 { 00:10:38.482 "name": "BaseBdev3", 00:10:38.482 "uuid": "a43fcab0-9dbf-4265-8bf6-6f0dc3568c79", 00:10:38.482 "is_configured": true, 00:10:38.482 "data_offset": 0, 00:10:38.482 "data_size": 65536 00:10:38.482 }, 00:10:38.482 { 00:10:38.482 "name": "BaseBdev4", 00:10:38.482 "uuid": "44b322df-c28c-48df-9239-87996deefbd5", 00:10:38.482 "is_configured": true, 00:10:38.482 "data_offset": 0, 00:10:38.482 "data_size": 65536 00:10:38.482 } 00:10:38.482 ] 00:10:38.482 } 00:10:38.482 } 00:10:38.482 }' 00:10:38.482 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:38.742 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:38.742 BaseBdev2 00:10:38.742 BaseBdev3 00:10:38.742 BaseBdev4' 00:10:38.742 10:59:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.742 [2024-10-29 10:59:44.224257] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:38.742 [2024-10-29 10:59:44.224400] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:38.742 [2024-10-29 10:59:44.224544] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:38.742 [2024-10-29 10:59:44.224700] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:38.742 [2024-10-29 10:59:44.224750] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 82328 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 82328 ']' 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 82328 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:10:38.742 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 82328 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 82328' 00:10:39.002 killing process with pid 82328 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 82328 00:10:39.002 [2024-10-29 10:59:44.275152] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:39.002 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 82328 00:10:39.002 [2024-10-29 10:59:44.352303] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:39.261 ************************************ 00:10:39.261 END TEST raid_state_function_test 00:10:39.261 ************************************ 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:39.261 00:10:39.261 real 0m9.563s 00:10:39.261 user 0m16.045s 00:10:39.261 sys 0m1.981s 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.261 10:59:44 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:10:39.261 10:59:44 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:39.261 10:59:44 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:39.261 10:59:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:39.261 ************************************ 00:10:39.261 START TEST raid_state_function_test_sb 00:10:39.261 ************************************ 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test concat 4 true 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:39.261 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:39.262 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:39.262 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:39.262 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:39.262 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:39.262 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82977 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82977' 00:10:39.522 Process raid pid: 82977 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82977 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 82977 ']' 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:39.522 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:39.522 10:59:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.522 [2024-10-29 10:59:44.846929] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:39.522 [2024-10-29 10:59:44.847156] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:39.522 [2024-10-29 10:59:45.000852] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:39.782 [2024-10-29 10:59:45.042341] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:39.782 [2024-10-29 10:59:45.119299] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:39.782 [2024-10-29 10:59:45.119449] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:40.351 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:40.351 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:10:40.351 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:40.351 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.351 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.352 [2024-10-29 10:59:45.707151] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:40.352 [2024-10-29 10:59:45.707314] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:40.352 [2024-10-29 10:59:45.707342] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:40.352 [2024-10-29 10:59:45.707373] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:40.352 [2024-10-29 10:59:45.707397] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:40.352 [2024-10-29 10:59:45.707438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:40.352 [2024-10-29 10:59:45.707483] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:40.352 [2024-10-29 10:59:45.707507] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.352 "name": "Existed_Raid", 00:10:40.352 "uuid": "cbcc92d2-6e85-4214-8503-63a0d5ae29da", 00:10:40.352 "strip_size_kb": 64, 00:10:40.352 "state": "configuring", 00:10:40.352 "raid_level": "concat", 00:10:40.352 "superblock": true, 00:10:40.352 "num_base_bdevs": 4, 00:10:40.352 "num_base_bdevs_discovered": 0, 00:10:40.352 "num_base_bdevs_operational": 4, 00:10:40.352 "base_bdevs_list": [ 00:10:40.352 { 00:10:40.352 "name": "BaseBdev1", 00:10:40.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.352 "is_configured": false, 00:10:40.352 "data_offset": 0, 00:10:40.352 "data_size": 0 00:10:40.352 }, 00:10:40.352 { 00:10:40.352 "name": "BaseBdev2", 00:10:40.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.352 "is_configured": false, 00:10:40.352 "data_offset": 0, 00:10:40.352 "data_size": 0 00:10:40.352 }, 00:10:40.352 { 00:10:40.352 "name": "BaseBdev3", 00:10:40.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.352 "is_configured": false, 00:10:40.352 "data_offset": 0, 00:10:40.352 "data_size": 0 00:10:40.352 }, 00:10:40.352 { 00:10:40.352 "name": "BaseBdev4", 00:10:40.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.352 "is_configured": false, 00:10:40.352 "data_offset": 0, 00:10:40.352 "data_size": 0 00:10:40.352 } 00:10:40.352 ] 00:10:40.352 }' 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.352 10:59:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 [2024-10-29 10:59:46.138298] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:40.922 [2024-10-29 10:59:46.138453] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 [2024-10-29 10:59:46.150256] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:40.922 [2024-10-29 10:59:46.150305] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:40.922 [2024-10-29 10:59:46.150313] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:40.922 [2024-10-29 10:59:46.150322] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:40.922 [2024-10-29 10:59:46.150328] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:40.922 [2024-10-29 10:59:46.150338] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:40.922 [2024-10-29 10:59:46.150344] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:40.922 [2024-10-29 10:59:46.150353] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 [2024-10-29 10:59:46.177246] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:40.922 BaseBdev1 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.922 [ 00:10:40.922 { 00:10:40.922 "name": "BaseBdev1", 00:10:40.922 "aliases": [ 00:10:40.922 "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9" 00:10:40.922 ], 00:10:40.922 "product_name": "Malloc disk", 00:10:40.922 "block_size": 512, 00:10:40.922 "num_blocks": 65536, 00:10:40.922 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:40.922 "assigned_rate_limits": { 00:10:40.922 "rw_ios_per_sec": 0, 00:10:40.922 "rw_mbytes_per_sec": 0, 00:10:40.922 "r_mbytes_per_sec": 0, 00:10:40.922 "w_mbytes_per_sec": 0 00:10:40.922 }, 00:10:40.922 "claimed": true, 00:10:40.922 "claim_type": "exclusive_write", 00:10:40.922 "zoned": false, 00:10:40.922 "supported_io_types": { 00:10:40.922 "read": true, 00:10:40.922 "write": true, 00:10:40.922 "unmap": true, 00:10:40.922 "flush": true, 00:10:40.922 "reset": true, 00:10:40.922 "nvme_admin": false, 00:10:40.922 "nvme_io": false, 00:10:40.922 "nvme_io_md": false, 00:10:40.922 "write_zeroes": true, 00:10:40.922 "zcopy": true, 00:10:40.922 "get_zone_info": false, 00:10:40.922 "zone_management": false, 00:10:40.922 "zone_append": false, 00:10:40.922 "compare": false, 00:10:40.922 "compare_and_write": false, 00:10:40.922 "abort": true, 00:10:40.922 "seek_hole": false, 00:10:40.922 "seek_data": false, 00:10:40.922 "copy": true, 00:10:40.922 "nvme_iov_md": false 00:10:40.922 }, 00:10:40.922 "memory_domains": [ 00:10:40.922 { 00:10:40.922 "dma_device_id": "system", 00:10:40.922 "dma_device_type": 1 00:10:40.922 }, 00:10:40.922 { 00:10:40.922 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.922 "dma_device_type": 2 00:10:40.922 } 00:10:40.922 ], 00:10:40.922 "driver_specific": {} 00:10:40.922 } 00:10:40.922 ] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.922 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.923 "name": "Existed_Raid", 00:10:40.923 "uuid": "4b305cb0-4cb2-4a7c-a537-59aea01a868d", 00:10:40.923 "strip_size_kb": 64, 00:10:40.923 "state": "configuring", 00:10:40.923 "raid_level": "concat", 00:10:40.923 "superblock": true, 00:10:40.923 "num_base_bdevs": 4, 00:10:40.923 "num_base_bdevs_discovered": 1, 00:10:40.923 "num_base_bdevs_operational": 4, 00:10:40.923 "base_bdevs_list": [ 00:10:40.923 { 00:10:40.923 "name": "BaseBdev1", 00:10:40.923 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:40.923 "is_configured": true, 00:10:40.923 "data_offset": 2048, 00:10:40.923 "data_size": 63488 00:10:40.923 }, 00:10:40.923 { 00:10:40.923 "name": "BaseBdev2", 00:10:40.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.923 "is_configured": false, 00:10:40.923 "data_offset": 0, 00:10:40.923 "data_size": 0 00:10:40.923 }, 00:10:40.923 { 00:10:40.923 "name": "BaseBdev3", 00:10:40.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.923 "is_configured": false, 00:10:40.923 "data_offset": 0, 00:10:40.923 "data_size": 0 00:10:40.923 }, 00:10:40.923 { 00:10:40.923 "name": "BaseBdev4", 00:10:40.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.923 "is_configured": false, 00:10:40.923 "data_offset": 0, 00:10:40.923 "data_size": 0 00:10:40.923 } 00:10:40.923 ] 00:10:40.923 }' 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.923 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.182 [2024-10-29 10:59:46.620582] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:41.182 [2024-10-29 10:59:46.620719] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.182 [2024-10-29 10:59:46.632577] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:41.182 [2024-10-29 10:59:46.634818] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:41.182 [2024-10-29 10:59:46.634863] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:41.182 [2024-10-29 10:59:46.634872] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:41.182 [2024-10-29 10:59:46.634881] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:41.182 [2024-10-29 10:59:46.634888] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:41.182 [2024-10-29 10:59:46.634896] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:41.182 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.183 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.442 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.442 "name": "Existed_Raid", 00:10:41.442 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:41.442 "strip_size_kb": 64, 00:10:41.442 "state": "configuring", 00:10:41.442 "raid_level": "concat", 00:10:41.442 "superblock": true, 00:10:41.442 "num_base_bdevs": 4, 00:10:41.442 "num_base_bdevs_discovered": 1, 00:10:41.442 "num_base_bdevs_operational": 4, 00:10:41.442 "base_bdevs_list": [ 00:10:41.442 { 00:10:41.442 "name": "BaseBdev1", 00:10:41.442 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:41.442 "is_configured": true, 00:10:41.442 "data_offset": 2048, 00:10:41.442 "data_size": 63488 00:10:41.442 }, 00:10:41.442 { 00:10:41.442 "name": "BaseBdev2", 00:10:41.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.442 "is_configured": false, 00:10:41.443 "data_offset": 0, 00:10:41.443 "data_size": 0 00:10:41.443 }, 00:10:41.443 { 00:10:41.443 "name": "BaseBdev3", 00:10:41.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.443 "is_configured": false, 00:10:41.443 "data_offset": 0, 00:10:41.443 "data_size": 0 00:10:41.443 }, 00:10:41.443 { 00:10:41.443 "name": "BaseBdev4", 00:10:41.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.443 "is_configured": false, 00:10:41.443 "data_offset": 0, 00:10:41.443 "data_size": 0 00:10:41.443 } 00:10:41.443 ] 00:10:41.443 }' 00:10:41.443 10:59:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.443 10:59:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.702 [2024-10-29 10:59:47.056806] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:41.702 BaseBdev2 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.702 [ 00:10:41.702 { 00:10:41.702 "name": "BaseBdev2", 00:10:41.702 "aliases": [ 00:10:41.702 "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9" 00:10:41.702 ], 00:10:41.702 "product_name": "Malloc disk", 00:10:41.702 "block_size": 512, 00:10:41.702 "num_blocks": 65536, 00:10:41.702 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:41.702 "assigned_rate_limits": { 00:10:41.702 "rw_ios_per_sec": 0, 00:10:41.702 "rw_mbytes_per_sec": 0, 00:10:41.702 "r_mbytes_per_sec": 0, 00:10:41.702 "w_mbytes_per_sec": 0 00:10:41.702 }, 00:10:41.702 "claimed": true, 00:10:41.702 "claim_type": "exclusive_write", 00:10:41.702 "zoned": false, 00:10:41.702 "supported_io_types": { 00:10:41.702 "read": true, 00:10:41.702 "write": true, 00:10:41.702 "unmap": true, 00:10:41.702 "flush": true, 00:10:41.702 "reset": true, 00:10:41.702 "nvme_admin": false, 00:10:41.702 "nvme_io": false, 00:10:41.702 "nvme_io_md": false, 00:10:41.702 "write_zeroes": true, 00:10:41.702 "zcopy": true, 00:10:41.702 "get_zone_info": false, 00:10:41.702 "zone_management": false, 00:10:41.702 "zone_append": false, 00:10:41.702 "compare": false, 00:10:41.702 "compare_and_write": false, 00:10:41.702 "abort": true, 00:10:41.702 "seek_hole": false, 00:10:41.702 "seek_data": false, 00:10:41.702 "copy": true, 00:10:41.702 "nvme_iov_md": false 00:10:41.702 }, 00:10:41.702 "memory_domains": [ 00:10:41.702 { 00:10:41.702 "dma_device_id": "system", 00:10:41.702 "dma_device_type": 1 00:10:41.702 }, 00:10:41.702 { 00:10:41.702 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.702 "dma_device_type": 2 00:10:41.702 } 00:10:41.702 ], 00:10:41.702 "driver_specific": {} 00:10:41.702 } 00:10:41.702 ] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.702 "name": "Existed_Raid", 00:10:41.702 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:41.702 "strip_size_kb": 64, 00:10:41.702 "state": "configuring", 00:10:41.702 "raid_level": "concat", 00:10:41.702 "superblock": true, 00:10:41.702 "num_base_bdevs": 4, 00:10:41.702 "num_base_bdevs_discovered": 2, 00:10:41.702 "num_base_bdevs_operational": 4, 00:10:41.702 "base_bdevs_list": [ 00:10:41.702 { 00:10:41.702 "name": "BaseBdev1", 00:10:41.702 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:41.702 "is_configured": true, 00:10:41.702 "data_offset": 2048, 00:10:41.702 "data_size": 63488 00:10:41.702 }, 00:10:41.702 { 00:10:41.702 "name": "BaseBdev2", 00:10:41.702 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:41.702 "is_configured": true, 00:10:41.702 "data_offset": 2048, 00:10:41.702 "data_size": 63488 00:10:41.702 }, 00:10:41.702 { 00:10:41.702 "name": "BaseBdev3", 00:10:41.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.702 "is_configured": false, 00:10:41.702 "data_offset": 0, 00:10:41.702 "data_size": 0 00:10:41.702 }, 00:10:41.702 { 00:10:41.702 "name": "BaseBdev4", 00:10:41.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.702 "is_configured": false, 00:10:41.702 "data_offset": 0, 00:10:41.702 "data_size": 0 00:10:41.702 } 00:10:41.702 ] 00:10:41.702 }' 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.702 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.273 [2024-10-29 10:59:47.532569] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:42.273 BaseBdev3 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.273 [ 00:10:42.273 { 00:10:42.273 "name": "BaseBdev3", 00:10:42.273 "aliases": [ 00:10:42.273 "02cd7117-0944-4e17-9c08-265893f86077" 00:10:42.273 ], 00:10:42.273 "product_name": "Malloc disk", 00:10:42.273 "block_size": 512, 00:10:42.273 "num_blocks": 65536, 00:10:42.273 "uuid": "02cd7117-0944-4e17-9c08-265893f86077", 00:10:42.273 "assigned_rate_limits": { 00:10:42.273 "rw_ios_per_sec": 0, 00:10:42.273 "rw_mbytes_per_sec": 0, 00:10:42.273 "r_mbytes_per_sec": 0, 00:10:42.273 "w_mbytes_per_sec": 0 00:10:42.273 }, 00:10:42.273 "claimed": true, 00:10:42.273 "claim_type": "exclusive_write", 00:10:42.273 "zoned": false, 00:10:42.273 "supported_io_types": { 00:10:42.273 "read": true, 00:10:42.273 "write": true, 00:10:42.273 "unmap": true, 00:10:42.273 "flush": true, 00:10:42.273 "reset": true, 00:10:42.273 "nvme_admin": false, 00:10:42.273 "nvme_io": false, 00:10:42.273 "nvme_io_md": false, 00:10:42.273 "write_zeroes": true, 00:10:42.273 "zcopy": true, 00:10:42.273 "get_zone_info": false, 00:10:42.273 "zone_management": false, 00:10:42.273 "zone_append": false, 00:10:42.273 "compare": false, 00:10:42.273 "compare_and_write": false, 00:10:42.273 "abort": true, 00:10:42.273 "seek_hole": false, 00:10:42.273 "seek_data": false, 00:10:42.273 "copy": true, 00:10:42.273 "nvme_iov_md": false 00:10:42.273 }, 00:10:42.273 "memory_domains": [ 00:10:42.273 { 00:10:42.273 "dma_device_id": "system", 00:10:42.273 "dma_device_type": 1 00:10:42.273 }, 00:10:42.273 { 00:10:42.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.273 "dma_device_type": 2 00:10:42.273 } 00:10:42.273 ], 00:10:42.273 "driver_specific": {} 00:10:42.273 } 00:10:42.273 ] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.273 "name": "Existed_Raid", 00:10:42.273 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:42.273 "strip_size_kb": 64, 00:10:42.273 "state": "configuring", 00:10:42.273 "raid_level": "concat", 00:10:42.273 "superblock": true, 00:10:42.273 "num_base_bdevs": 4, 00:10:42.273 "num_base_bdevs_discovered": 3, 00:10:42.273 "num_base_bdevs_operational": 4, 00:10:42.273 "base_bdevs_list": [ 00:10:42.273 { 00:10:42.273 "name": "BaseBdev1", 00:10:42.273 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:42.273 "is_configured": true, 00:10:42.273 "data_offset": 2048, 00:10:42.273 "data_size": 63488 00:10:42.273 }, 00:10:42.273 { 00:10:42.273 "name": "BaseBdev2", 00:10:42.273 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:42.273 "is_configured": true, 00:10:42.273 "data_offset": 2048, 00:10:42.273 "data_size": 63488 00:10:42.273 }, 00:10:42.273 { 00:10:42.273 "name": "BaseBdev3", 00:10:42.273 "uuid": "02cd7117-0944-4e17-9c08-265893f86077", 00:10:42.273 "is_configured": true, 00:10:42.273 "data_offset": 2048, 00:10:42.273 "data_size": 63488 00:10:42.273 }, 00:10:42.273 { 00:10:42.273 "name": "BaseBdev4", 00:10:42.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.273 "is_configured": false, 00:10:42.273 "data_offset": 0, 00:10:42.273 "data_size": 0 00:10:42.273 } 00:10:42.273 ] 00:10:42.273 }' 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.273 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.534 [2024-10-29 10:59:47.988920] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:42.534 [2024-10-29 10:59:47.989303] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:42.534 [2024-10-29 10:59:47.989377] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:42.534 BaseBdev4 00:10:42.534 [2024-10-29 10:59:47.989734] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:42.534 [2024-10-29 10:59:47.989882] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:42.534 [2024-10-29 10:59:47.989896] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:10:42.534 [2024-10-29 10:59:47.990037] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.534 10:59:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.534 [ 00:10:42.534 { 00:10:42.534 "name": "BaseBdev4", 00:10:42.534 "aliases": [ 00:10:42.534 "8af11bd5-d829-401e-a68f-0dc16989b623" 00:10:42.534 ], 00:10:42.534 "product_name": "Malloc disk", 00:10:42.534 "block_size": 512, 00:10:42.534 "num_blocks": 65536, 00:10:42.534 "uuid": "8af11bd5-d829-401e-a68f-0dc16989b623", 00:10:42.534 "assigned_rate_limits": { 00:10:42.534 "rw_ios_per_sec": 0, 00:10:42.534 "rw_mbytes_per_sec": 0, 00:10:42.534 "r_mbytes_per_sec": 0, 00:10:42.534 "w_mbytes_per_sec": 0 00:10:42.534 }, 00:10:42.534 "claimed": true, 00:10:42.534 "claim_type": "exclusive_write", 00:10:42.534 "zoned": false, 00:10:42.534 "supported_io_types": { 00:10:42.534 "read": true, 00:10:42.534 "write": true, 00:10:42.534 "unmap": true, 00:10:42.534 "flush": true, 00:10:42.534 "reset": true, 00:10:42.534 "nvme_admin": false, 00:10:42.534 "nvme_io": false, 00:10:42.534 "nvme_io_md": false, 00:10:42.534 "write_zeroes": true, 00:10:42.534 "zcopy": true, 00:10:42.534 "get_zone_info": false, 00:10:42.534 "zone_management": false, 00:10:42.534 "zone_append": false, 00:10:42.534 "compare": false, 00:10:42.534 "compare_and_write": false, 00:10:42.534 "abort": true, 00:10:42.534 "seek_hole": false, 00:10:42.534 "seek_data": false, 00:10:42.534 "copy": true, 00:10:42.534 "nvme_iov_md": false 00:10:42.534 }, 00:10:42.534 "memory_domains": [ 00:10:42.534 { 00:10:42.534 "dma_device_id": "system", 00:10:42.534 "dma_device_type": 1 00:10:42.534 }, 00:10:42.534 { 00:10:42.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.534 "dma_device_type": 2 00:10:42.534 } 00:10:42.534 ], 00:10:42.534 "driver_specific": {} 00:10:42.534 } 00:10:42.534 ] 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.534 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.795 "name": "Existed_Raid", 00:10:42.795 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:42.795 "strip_size_kb": 64, 00:10:42.795 "state": "online", 00:10:42.795 "raid_level": "concat", 00:10:42.795 "superblock": true, 00:10:42.795 "num_base_bdevs": 4, 00:10:42.795 "num_base_bdevs_discovered": 4, 00:10:42.795 "num_base_bdevs_operational": 4, 00:10:42.795 "base_bdevs_list": [ 00:10:42.795 { 00:10:42.795 "name": "BaseBdev1", 00:10:42.795 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:42.795 "is_configured": true, 00:10:42.795 "data_offset": 2048, 00:10:42.795 "data_size": 63488 00:10:42.795 }, 00:10:42.795 { 00:10:42.795 "name": "BaseBdev2", 00:10:42.795 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:42.795 "is_configured": true, 00:10:42.795 "data_offset": 2048, 00:10:42.795 "data_size": 63488 00:10:42.795 }, 00:10:42.795 { 00:10:42.795 "name": "BaseBdev3", 00:10:42.795 "uuid": "02cd7117-0944-4e17-9c08-265893f86077", 00:10:42.795 "is_configured": true, 00:10:42.795 "data_offset": 2048, 00:10:42.795 "data_size": 63488 00:10:42.795 }, 00:10:42.795 { 00:10:42.795 "name": "BaseBdev4", 00:10:42.795 "uuid": "8af11bd5-d829-401e-a68f-0dc16989b623", 00:10:42.795 "is_configured": true, 00:10:42.795 "data_offset": 2048, 00:10:42.795 "data_size": 63488 00:10:42.795 } 00:10:42.795 ] 00:10:42.795 }' 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.795 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.055 [2024-10-29 10:59:48.412749] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.055 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:43.055 "name": "Existed_Raid", 00:10:43.055 "aliases": [ 00:10:43.055 "367560ae-f8cf-45b6-a981-0f9398501893" 00:10:43.055 ], 00:10:43.055 "product_name": "Raid Volume", 00:10:43.055 "block_size": 512, 00:10:43.055 "num_blocks": 253952, 00:10:43.055 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:43.055 "assigned_rate_limits": { 00:10:43.055 "rw_ios_per_sec": 0, 00:10:43.055 "rw_mbytes_per_sec": 0, 00:10:43.055 "r_mbytes_per_sec": 0, 00:10:43.055 "w_mbytes_per_sec": 0 00:10:43.055 }, 00:10:43.055 "claimed": false, 00:10:43.055 "zoned": false, 00:10:43.055 "supported_io_types": { 00:10:43.055 "read": true, 00:10:43.055 "write": true, 00:10:43.055 "unmap": true, 00:10:43.055 "flush": true, 00:10:43.055 "reset": true, 00:10:43.055 "nvme_admin": false, 00:10:43.055 "nvme_io": false, 00:10:43.055 "nvme_io_md": false, 00:10:43.055 "write_zeroes": true, 00:10:43.055 "zcopy": false, 00:10:43.056 "get_zone_info": false, 00:10:43.056 "zone_management": false, 00:10:43.056 "zone_append": false, 00:10:43.056 "compare": false, 00:10:43.056 "compare_and_write": false, 00:10:43.056 "abort": false, 00:10:43.056 "seek_hole": false, 00:10:43.056 "seek_data": false, 00:10:43.056 "copy": false, 00:10:43.056 "nvme_iov_md": false 00:10:43.056 }, 00:10:43.056 "memory_domains": [ 00:10:43.056 { 00:10:43.056 "dma_device_id": "system", 00:10:43.056 "dma_device_type": 1 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.056 "dma_device_type": 2 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "system", 00:10:43.056 "dma_device_type": 1 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.056 "dma_device_type": 2 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "system", 00:10:43.056 "dma_device_type": 1 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.056 "dma_device_type": 2 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "system", 00:10:43.056 "dma_device_type": 1 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.056 "dma_device_type": 2 00:10:43.056 } 00:10:43.056 ], 00:10:43.056 "driver_specific": { 00:10:43.056 "raid": { 00:10:43.056 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:43.056 "strip_size_kb": 64, 00:10:43.056 "state": "online", 00:10:43.056 "raid_level": "concat", 00:10:43.056 "superblock": true, 00:10:43.056 "num_base_bdevs": 4, 00:10:43.056 "num_base_bdevs_discovered": 4, 00:10:43.056 "num_base_bdevs_operational": 4, 00:10:43.056 "base_bdevs_list": [ 00:10:43.056 { 00:10:43.056 "name": "BaseBdev1", 00:10:43.056 "uuid": "b3e6b7e8-24f6-49db-84e4-ac69a9d724d9", 00:10:43.056 "is_configured": true, 00:10:43.056 "data_offset": 2048, 00:10:43.056 "data_size": 63488 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "name": "BaseBdev2", 00:10:43.056 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:43.056 "is_configured": true, 00:10:43.056 "data_offset": 2048, 00:10:43.056 "data_size": 63488 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "name": "BaseBdev3", 00:10:43.056 "uuid": "02cd7117-0944-4e17-9c08-265893f86077", 00:10:43.056 "is_configured": true, 00:10:43.056 "data_offset": 2048, 00:10:43.056 "data_size": 63488 00:10:43.056 }, 00:10:43.056 { 00:10:43.056 "name": "BaseBdev4", 00:10:43.056 "uuid": "8af11bd5-d829-401e-a68f-0dc16989b623", 00:10:43.056 "is_configured": true, 00:10:43.056 "data_offset": 2048, 00:10:43.056 "data_size": 63488 00:10:43.056 } 00:10:43.056 ] 00:10:43.056 } 00:10:43.056 } 00:10:43.056 }' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:43.056 BaseBdev2 00:10:43.056 BaseBdev3 00:10:43.056 BaseBdev4' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.056 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.316 [2024-10-29 10:59:48.707906] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:43.316 [2024-10-29 10:59:48.708034] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:43.316 [2024-10-29 10:59:48.708132] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.316 "name": "Existed_Raid", 00:10:43.316 "uuid": "367560ae-f8cf-45b6-a981-0f9398501893", 00:10:43.316 "strip_size_kb": 64, 00:10:43.316 "state": "offline", 00:10:43.316 "raid_level": "concat", 00:10:43.316 "superblock": true, 00:10:43.316 "num_base_bdevs": 4, 00:10:43.316 "num_base_bdevs_discovered": 3, 00:10:43.316 "num_base_bdevs_operational": 3, 00:10:43.316 "base_bdevs_list": [ 00:10:43.316 { 00:10:43.316 "name": null, 00:10:43.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.316 "is_configured": false, 00:10:43.316 "data_offset": 0, 00:10:43.316 "data_size": 63488 00:10:43.316 }, 00:10:43.316 { 00:10:43.316 "name": "BaseBdev2", 00:10:43.316 "uuid": "2076b3bc-ca92-417e-90a3-3dfa0cc8b3b9", 00:10:43.316 "is_configured": true, 00:10:43.316 "data_offset": 2048, 00:10:43.316 "data_size": 63488 00:10:43.316 }, 00:10:43.316 { 00:10:43.316 "name": "BaseBdev3", 00:10:43.316 "uuid": "02cd7117-0944-4e17-9c08-265893f86077", 00:10:43.316 "is_configured": true, 00:10:43.316 "data_offset": 2048, 00:10:43.316 "data_size": 63488 00:10:43.316 }, 00:10:43.316 { 00:10:43.316 "name": "BaseBdev4", 00:10:43.316 "uuid": "8af11bd5-d829-401e-a68f-0dc16989b623", 00:10:43.316 "is_configured": true, 00:10:43.316 "data_offset": 2048, 00:10:43.316 "data_size": 63488 00:10:43.316 } 00:10:43.316 ] 00:10:43.316 }' 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.316 10:59:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 [2024-10-29 10:59:49.215650] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 [2024-10-29 10:59:49.292251] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.886 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.886 [2024-10-29 10:59:49.372680] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:43.886 [2024-10-29 10:59:49.372819] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 BaseBdev2 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 [ 00:10:44.179 { 00:10:44.179 "name": "BaseBdev2", 00:10:44.179 "aliases": [ 00:10:44.179 "2d677b77-afe1-46a3-8141-46bf8ce017a1" 00:10:44.179 ], 00:10:44.179 "product_name": "Malloc disk", 00:10:44.179 "block_size": 512, 00:10:44.179 "num_blocks": 65536, 00:10:44.179 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:44.179 "assigned_rate_limits": { 00:10:44.179 "rw_ios_per_sec": 0, 00:10:44.179 "rw_mbytes_per_sec": 0, 00:10:44.179 "r_mbytes_per_sec": 0, 00:10:44.179 "w_mbytes_per_sec": 0 00:10:44.179 }, 00:10:44.179 "claimed": false, 00:10:44.179 "zoned": false, 00:10:44.179 "supported_io_types": { 00:10:44.179 "read": true, 00:10:44.179 "write": true, 00:10:44.179 "unmap": true, 00:10:44.179 "flush": true, 00:10:44.179 "reset": true, 00:10:44.179 "nvme_admin": false, 00:10:44.179 "nvme_io": false, 00:10:44.179 "nvme_io_md": false, 00:10:44.179 "write_zeroes": true, 00:10:44.179 "zcopy": true, 00:10:44.179 "get_zone_info": false, 00:10:44.179 "zone_management": false, 00:10:44.179 "zone_append": false, 00:10:44.179 "compare": false, 00:10:44.179 "compare_and_write": false, 00:10:44.179 "abort": true, 00:10:44.179 "seek_hole": false, 00:10:44.179 "seek_data": false, 00:10:44.179 "copy": true, 00:10:44.179 "nvme_iov_md": false 00:10:44.179 }, 00:10:44.179 "memory_domains": [ 00:10:44.179 { 00:10:44.179 "dma_device_id": "system", 00:10:44.179 "dma_device_type": 1 00:10:44.179 }, 00:10:44.179 { 00:10:44.179 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:44.179 "dma_device_type": 2 00:10:44.179 } 00:10:44.179 ], 00:10:44.179 "driver_specific": {} 00:10:44.179 } 00:10:44.179 ] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 BaseBdev3 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 [ 00:10:44.179 { 00:10:44.179 "name": "BaseBdev3", 00:10:44.179 "aliases": [ 00:10:44.179 "2db5e54c-06e1-4c35-a929-b2206a81b5df" 00:10:44.179 ], 00:10:44.179 "product_name": "Malloc disk", 00:10:44.179 "block_size": 512, 00:10:44.179 "num_blocks": 65536, 00:10:44.179 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:44.179 "assigned_rate_limits": { 00:10:44.179 "rw_ios_per_sec": 0, 00:10:44.179 "rw_mbytes_per_sec": 0, 00:10:44.179 "r_mbytes_per_sec": 0, 00:10:44.179 "w_mbytes_per_sec": 0 00:10:44.179 }, 00:10:44.179 "claimed": false, 00:10:44.179 "zoned": false, 00:10:44.179 "supported_io_types": { 00:10:44.179 "read": true, 00:10:44.179 "write": true, 00:10:44.179 "unmap": true, 00:10:44.179 "flush": true, 00:10:44.179 "reset": true, 00:10:44.179 "nvme_admin": false, 00:10:44.179 "nvme_io": false, 00:10:44.179 "nvme_io_md": false, 00:10:44.179 "write_zeroes": true, 00:10:44.179 "zcopy": true, 00:10:44.179 "get_zone_info": false, 00:10:44.179 "zone_management": false, 00:10:44.179 "zone_append": false, 00:10:44.179 "compare": false, 00:10:44.179 "compare_and_write": false, 00:10:44.179 "abort": true, 00:10:44.179 "seek_hole": false, 00:10:44.179 "seek_data": false, 00:10:44.179 "copy": true, 00:10:44.179 "nvme_iov_md": false 00:10:44.179 }, 00:10:44.179 "memory_domains": [ 00:10:44.179 { 00:10:44.179 "dma_device_id": "system", 00:10:44.179 "dma_device_type": 1 00:10:44.179 }, 00:10:44.179 { 00:10:44.179 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:44.179 "dma_device_type": 2 00:10:44.179 } 00:10:44.179 ], 00:10:44.179 "driver_specific": {} 00:10:44.179 } 00:10:44.179 ] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.179 BaseBdev4 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:44.179 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.180 [ 00:10:44.180 { 00:10:44.180 "name": "BaseBdev4", 00:10:44.180 "aliases": [ 00:10:44.180 "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3" 00:10:44.180 ], 00:10:44.180 "product_name": "Malloc disk", 00:10:44.180 "block_size": 512, 00:10:44.180 "num_blocks": 65536, 00:10:44.180 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:44.180 "assigned_rate_limits": { 00:10:44.180 "rw_ios_per_sec": 0, 00:10:44.180 "rw_mbytes_per_sec": 0, 00:10:44.180 "r_mbytes_per_sec": 0, 00:10:44.180 "w_mbytes_per_sec": 0 00:10:44.180 }, 00:10:44.180 "claimed": false, 00:10:44.180 "zoned": false, 00:10:44.180 "supported_io_types": { 00:10:44.180 "read": true, 00:10:44.180 "write": true, 00:10:44.180 "unmap": true, 00:10:44.180 "flush": true, 00:10:44.180 "reset": true, 00:10:44.180 "nvme_admin": false, 00:10:44.180 "nvme_io": false, 00:10:44.180 "nvme_io_md": false, 00:10:44.180 "write_zeroes": true, 00:10:44.180 "zcopy": true, 00:10:44.180 "get_zone_info": false, 00:10:44.180 "zone_management": false, 00:10:44.180 "zone_append": false, 00:10:44.180 "compare": false, 00:10:44.180 "compare_and_write": false, 00:10:44.180 "abort": true, 00:10:44.180 "seek_hole": false, 00:10:44.180 "seek_data": false, 00:10:44.180 "copy": true, 00:10:44.180 "nvme_iov_md": false 00:10:44.180 }, 00:10:44.180 "memory_domains": [ 00:10:44.180 { 00:10:44.180 "dma_device_id": "system", 00:10:44.180 "dma_device_type": 1 00:10:44.180 }, 00:10:44.180 { 00:10:44.180 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:44.180 "dma_device_type": 2 00:10:44.180 } 00:10:44.180 ], 00:10:44.180 "driver_specific": {} 00:10:44.180 } 00:10:44.180 ] 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.180 [2024-10-29 10:59:49.610030] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:44.180 [2024-10-29 10:59:49.610156] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:44.180 [2024-10-29 10:59:49.610198] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:44.180 [2024-10-29 10:59:49.612398] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:44.180 [2024-10-29 10:59:49.612491] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.180 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.476 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.476 "name": "Existed_Raid", 00:10:44.476 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:44.476 "strip_size_kb": 64, 00:10:44.476 "state": "configuring", 00:10:44.476 "raid_level": "concat", 00:10:44.476 "superblock": true, 00:10:44.476 "num_base_bdevs": 4, 00:10:44.476 "num_base_bdevs_discovered": 3, 00:10:44.476 "num_base_bdevs_operational": 4, 00:10:44.476 "base_bdevs_list": [ 00:10:44.476 { 00:10:44.476 "name": "BaseBdev1", 00:10:44.476 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.476 "is_configured": false, 00:10:44.476 "data_offset": 0, 00:10:44.476 "data_size": 0 00:10:44.476 }, 00:10:44.476 { 00:10:44.476 "name": "BaseBdev2", 00:10:44.476 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:44.476 "is_configured": true, 00:10:44.476 "data_offset": 2048, 00:10:44.476 "data_size": 63488 00:10:44.476 }, 00:10:44.476 { 00:10:44.476 "name": "BaseBdev3", 00:10:44.476 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:44.476 "is_configured": true, 00:10:44.476 "data_offset": 2048, 00:10:44.476 "data_size": 63488 00:10:44.476 }, 00:10:44.476 { 00:10:44.476 "name": "BaseBdev4", 00:10:44.476 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:44.476 "is_configured": true, 00:10:44.476 "data_offset": 2048, 00:10:44.476 "data_size": 63488 00:10:44.476 } 00:10:44.476 ] 00:10:44.476 }' 00:10:44.476 10:59:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.476 10:59:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.735 [2024-10-29 10:59:50.089272] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.735 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.735 "name": "Existed_Raid", 00:10:44.735 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:44.735 "strip_size_kb": 64, 00:10:44.735 "state": "configuring", 00:10:44.735 "raid_level": "concat", 00:10:44.735 "superblock": true, 00:10:44.735 "num_base_bdevs": 4, 00:10:44.735 "num_base_bdevs_discovered": 2, 00:10:44.735 "num_base_bdevs_operational": 4, 00:10:44.735 "base_bdevs_list": [ 00:10:44.735 { 00:10:44.735 "name": "BaseBdev1", 00:10:44.735 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.735 "is_configured": false, 00:10:44.735 "data_offset": 0, 00:10:44.735 "data_size": 0 00:10:44.735 }, 00:10:44.735 { 00:10:44.735 "name": null, 00:10:44.735 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:44.735 "is_configured": false, 00:10:44.735 "data_offset": 0, 00:10:44.735 "data_size": 63488 00:10:44.735 }, 00:10:44.735 { 00:10:44.735 "name": "BaseBdev3", 00:10:44.735 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:44.735 "is_configured": true, 00:10:44.735 "data_offset": 2048, 00:10:44.735 "data_size": 63488 00:10:44.735 }, 00:10:44.735 { 00:10:44.736 "name": "BaseBdev4", 00:10:44.736 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:44.736 "is_configured": true, 00:10:44.736 "data_offset": 2048, 00:10:44.736 "data_size": 63488 00:10:44.736 } 00:10:44.736 ] 00:10:44.736 }' 00:10:44.736 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.736 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.302 [2024-10-29 10:59:50.553409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:45.302 BaseBdev1 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.302 [ 00:10:45.302 { 00:10:45.302 "name": "BaseBdev1", 00:10:45.302 "aliases": [ 00:10:45.302 "c35b3820-177a-4e5a-89d8-813b455d6e2a" 00:10:45.302 ], 00:10:45.302 "product_name": "Malloc disk", 00:10:45.302 "block_size": 512, 00:10:45.302 "num_blocks": 65536, 00:10:45.302 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:45.302 "assigned_rate_limits": { 00:10:45.302 "rw_ios_per_sec": 0, 00:10:45.302 "rw_mbytes_per_sec": 0, 00:10:45.302 "r_mbytes_per_sec": 0, 00:10:45.302 "w_mbytes_per_sec": 0 00:10:45.302 }, 00:10:45.302 "claimed": true, 00:10:45.302 "claim_type": "exclusive_write", 00:10:45.302 "zoned": false, 00:10:45.302 "supported_io_types": { 00:10:45.302 "read": true, 00:10:45.302 "write": true, 00:10:45.302 "unmap": true, 00:10:45.302 "flush": true, 00:10:45.302 "reset": true, 00:10:45.302 "nvme_admin": false, 00:10:45.302 "nvme_io": false, 00:10:45.302 "nvme_io_md": false, 00:10:45.302 "write_zeroes": true, 00:10:45.302 "zcopy": true, 00:10:45.302 "get_zone_info": false, 00:10:45.302 "zone_management": false, 00:10:45.302 "zone_append": false, 00:10:45.302 "compare": false, 00:10:45.302 "compare_and_write": false, 00:10:45.302 "abort": true, 00:10:45.302 "seek_hole": false, 00:10:45.302 "seek_data": false, 00:10:45.302 "copy": true, 00:10:45.302 "nvme_iov_md": false 00:10:45.302 }, 00:10:45.302 "memory_domains": [ 00:10:45.302 { 00:10:45.302 "dma_device_id": "system", 00:10:45.302 "dma_device_type": 1 00:10:45.302 }, 00:10:45.302 { 00:10:45.302 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:45.302 "dma_device_type": 2 00:10:45.302 } 00:10:45.302 ], 00:10:45.302 "driver_specific": {} 00:10:45.302 } 00:10:45.302 ] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:45.302 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.303 "name": "Existed_Raid", 00:10:45.303 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:45.303 "strip_size_kb": 64, 00:10:45.303 "state": "configuring", 00:10:45.303 "raid_level": "concat", 00:10:45.303 "superblock": true, 00:10:45.303 "num_base_bdevs": 4, 00:10:45.303 "num_base_bdevs_discovered": 3, 00:10:45.303 "num_base_bdevs_operational": 4, 00:10:45.303 "base_bdevs_list": [ 00:10:45.303 { 00:10:45.303 "name": "BaseBdev1", 00:10:45.303 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:45.303 "is_configured": true, 00:10:45.303 "data_offset": 2048, 00:10:45.303 "data_size": 63488 00:10:45.303 }, 00:10:45.303 { 00:10:45.303 "name": null, 00:10:45.303 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:45.303 "is_configured": false, 00:10:45.303 "data_offset": 0, 00:10:45.303 "data_size": 63488 00:10:45.303 }, 00:10:45.303 { 00:10:45.303 "name": "BaseBdev3", 00:10:45.303 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:45.303 "is_configured": true, 00:10:45.303 "data_offset": 2048, 00:10:45.303 "data_size": 63488 00:10:45.303 }, 00:10:45.303 { 00:10:45.303 "name": "BaseBdev4", 00:10:45.303 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:45.303 "is_configured": true, 00:10:45.303 "data_offset": 2048, 00:10:45.303 "data_size": 63488 00:10:45.303 } 00:10:45.303 ] 00:10:45.303 }' 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.303 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.560 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.560 10:59:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:45.560 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.560 10:59:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.560 [2024-10-29 10:59:51.052653] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.560 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.818 "name": "Existed_Raid", 00:10:45.818 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:45.818 "strip_size_kb": 64, 00:10:45.818 "state": "configuring", 00:10:45.818 "raid_level": "concat", 00:10:45.818 "superblock": true, 00:10:45.818 "num_base_bdevs": 4, 00:10:45.818 "num_base_bdevs_discovered": 2, 00:10:45.818 "num_base_bdevs_operational": 4, 00:10:45.818 "base_bdevs_list": [ 00:10:45.818 { 00:10:45.818 "name": "BaseBdev1", 00:10:45.818 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:45.818 "is_configured": true, 00:10:45.818 "data_offset": 2048, 00:10:45.818 "data_size": 63488 00:10:45.818 }, 00:10:45.818 { 00:10:45.818 "name": null, 00:10:45.818 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:45.818 "is_configured": false, 00:10:45.818 "data_offset": 0, 00:10:45.818 "data_size": 63488 00:10:45.818 }, 00:10:45.818 { 00:10:45.818 "name": null, 00:10:45.818 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:45.818 "is_configured": false, 00:10:45.818 "data_offset": 0, 00:10:45.818 "data_size": 63488 00:10:45.818 }, 00:10:45.818 { 00:10:45.818 "name": "BaseBdev4", 00:10:45.818 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:45.818 "is_configured": true, 00:10:45.818 "data_offset": 2048, 00:10:45.818 "data_size": 63488 00:10:45.818 } 00:10:45.818 ] 00:10:45.818 }' 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.818 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.075 [2024-10-29 10:59:51.475960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:46.075 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.076 "name": "Existed_Raid", 00:10:46.076 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:46.076 "strip_size_kb": 64, 00:10:46.076 "state": "configuring", 00:10:46.076 "raid_level": "concat", 00:10:46.076 "superblock": true, 00:10:46.076 "num_base_bdevs": 4, 00:10:46.076 "num_base_bdevs_discovered": 3, 00:10:46.076 "num_base_bdevs_operational": 4, 00:10:46.076 "base_bdevs_list": [ 00:10:46.076 { 00:10:46.076 "name": "BaseBdev1", 00:10:46.076 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:46.076 "is_configured": true, 00:10:46.076 "data_offset": 2048, 00:10:46.076 "data_size": 63488 00:10:46.076 }, 00:10:46.076 { 00:10:46.076 "name": null, 00:10:46.076 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:46.076 "is_configured": false, 00:10:46.076 "data_offset": 0, 00:10:46.076 "data_size": 63488 00:10:46.076 }, 00:10:46.076 { 00:10:46.076 "name": "BaseBdev3", 00:10:46.076 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:46.076 "is_configured": true, 00:10:46.076 "data_offset": 2048, 00:10:46.076 "data_size": 63488 00:10:46.076 }, 00:10:46.076 { 00:10:46.076 "name": "BaseBdev4", 00:10:46.076 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:46.076 "is_configured": true, 00:10:46.076 "data_offset": 2048, 00:10:46.076 "data_size": 63488 00:10:46.076 } 00:10:46.076 ] 00:10:46.076 }' 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.076 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.641 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:46.641 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.641 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.641 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.642 [2024-10-29 10:59:51.951613] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.642 10:59:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.642 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.642 "name": "Existed_Raid", 00:10:46.642 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:46.642 "strip_size_kb": 64, 00:10:46.642 "state": "configuring", 00:10:46.642 "raid_level": "concat", 00:10:46.642 "superblock": true, 00:10:46.642 "num_base_bdevs": 4, 00:10:46.642 "num_base_bdevs_discovered": 2, 00:10:46.642 "num_base_bdevs_operational": 4, 00:10:46.642 "base_bdevs_list": [ 00:10:46.642 { 00:10:46.642 "name": null, 00:10:46.642 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:46.642 "is_configured": false, 00:10:46.642 "data_offset": 0, 00:10:46.642 "data_size": 63488 00:10:46.642 }, 00:10:46.642 { 00:10:46.642 "name": null, 00:10:46.642 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:46.642 "is_configured": false, 00:10:46.642 "data_offset": 0, 00:10:46.642 "data_size": 63488 00:10:46.642 }, 00:10:46.642 { 00:10:46.642 "name": "BaseBdev3", 00:10:46.642 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:46.642 "is_configured": true, 00:10:46.642 "data_offset": 2048, 00:10:46.642 "data_size": 63488 00:10:46.642 }, 00:10:46.642 { 00:10:46.642 "name": "BaseBdev4", 00:10:46.642 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:46.642 "is_configured": true, 00:10:46.642 "data_offset": 2048, 00:10:46.642 "data_size": 63488 00:10:46.642 } 00:10:46.642 ] 00:10:46.642 }' 00:10:46.642 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.642 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.209 [2024-10-29 10:59:52.450390] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.209 "name": "Existed_Raid", 00:10:47.209 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:47.209 "strip_size_kb": 64, 00:10:47.209 "state": "configuring", 00:10:47.209 "raid_level": "concat", 00:10:47.209 "superblock": true, 00:10:47.209 "num_base_bdevs": 4, 00:10:47.209 "num_base_bdevs_discovered": 3, 00:10:47.209 "num_base_bdevs_operational": 4, 00:10:47.209 "base_bdevs_list": [ 00:10:47.209 { 00:10:47.209 "name": null, 00:10:47.209 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:47.209 "is_configured": false, 00:10:47.209 "data_offset": 0, 00:10:47.209 "data_size": 63488 00:10:47.209 }, 00:10:47.209 { 00:10:47.209 "name": "BaseBdev2", 00:10:47.209 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:47.209 "is_configured": true, 00:10:47.209 "data_offset": 2048, 00:10:47.209 "data_size": 63488 00:10:47.209 }, 00:10:47.209 { 00:10:47.209 "name": "BaseBdev3", 00:10:47.209 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:47.209 "is_configured": true, 00:10:47.209 "data_offset": 2048, 00:10:47.209 "data_size": 63488 00:10:47.209 }, 00:10:47.209 { 00:10:47.209 "name": "BaseBdev4", 00:10:47.209 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:47.209 "is_configured": true, 00:10:47.209 "data_offset": 2048, 00:10:47.209 "data_size": 63488 00:10:47.209 } 00:10:47.209 ] 00:10:47.209 }' 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.209 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c35b3820-177a-4e5a-89d8-813b455d6e2a 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.467 [2024-10-29 10:59:52.958250] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:47.467 [2024-10-29 10:59:52.958567] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:10:47.467 [2024-10-29 10:59:52.958617] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:47.467 NewBaseBdev 00:10:47.467 [2024-10-29 10:59:52.958983] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:10:47.467 [2024-10-29 10:59:52.959117] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:10:47.467 [2024-10-29 10:59:52.959182] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:10:47.467 [2024-10-29 10:59:52.959341] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.467 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.726 [ 00:10:47.726 { 00:10:47.726 "name": "NewBaseBdev", 00:10:47.726 "aliases": [ 00:10:47.726 "c35b3820-177a-4e5a-89d8-813b455d6e2a" 00:10:47.726 ], 00:10:47.726 "product_name": "Malloc disk", 00:10:47.726 "block_size": 512, 00:10:47.726 "num_blocks": 65536, 00:10:47.726 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:47.726 "assigned_rate_limits": { 00:10:47.726 "rw_ios_per_sec": 0, 00:10:47.726 "rw_mbytes_per_sec": 0, 00:10:47.726 "r_mbytes_per_sec": 0, 00:10:47.726 "w_mbytes_per_sec": 0 00:10:47.726 }, 00:10:47.726 "claimed": true, 00:10:47.726 "claim_type": "exclusive_write", 00:10:47.726 "zoned": false, 00:10:47.726 "supported_io_types": { 00:10:47.726 "read": true, 00:10:47.726 "write": true, 00:10:47.726 "unmap": true, 00:10:47.726 "flush": true, 00:10:47.726 "reset": true, 00:10:47.726 "nvme_admin": false, 00:10:47.726 "nvme_io": false, 00:10:47.726 "nvme_io_md": false, 00:10:47.726 "write_zeroes": true, 00:10:47.726 "zcopy": true, 00:10:47.726 "get_zone_info": false, 00:10:47.726 "zone_management": false, 00:10:47.726 "zone_append": false, 00:10:47.726 "compare": false, 00:10:47.726 "compare_and_write": false, 00:10:47.726 "abort": true, 00:10:47.726 "seek_hole": false, 00:10:47.726 "seek_data": false, 00:10:47.726 "copy": true, 00:10:47.726 "nvme_iov_md": false 00:10:47.726 }, 00:10:47.726 "memory_domains": [ 00:10:47.726 { 00:10:47.726 "dma_device_id": "system", 00:10:47.726 "dma_device_type": 1 00:10:47.726 }, 00:10:47.726 { 00:10:47.726 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.726 "dma_device_type": 2 00:10:47.726 } 00:10:47.726 ], 00:10:47.726 "driver_specific": {} 00:10:47.726 } 00:10:47.726 ] 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:47.726 10:59:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.726 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:47.726 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.726 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.726 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.726 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.726 "name": "Existed_Raid", 00:10:47.726 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:47.726 "strip_size_kb": 64, 00:10:47.726 "state": "online", 00:10:47.726 "raid_level": "concat", 00:10:47.726 "superblock": true, 00:10:47.726 "num_base_bdevs": 4, 00:10:47.726 "num_base_bdevs_discovered": 4, 00:10:47.726 "num_base_bdevs_operational": 4, 00:10:47.726 "base_bdevs_list": [ 00:10:47.726 { 00:10:47.726 "name": "NewBaseBdev", 00:10:47.726 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:47.726 "is_configured": true, 00:10:47.726 "data_offset": 2048, 00:10:47.726 "data_size": 63488 00:10:47.726 }, 00:10:47.726 { 00:10:47.726 "name": "BaseBdev2", 00:10:47.726 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:47.726 "is_configured": true, 00:10:47.726 "data_offset": 2048, 00:10:47.726 "data_size": 63488 00:10:47.726 }, 00:10:47.726 { 00:10:47.726 "name": "BaseBdev3", 00:10:47.726 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:47.726 "is_configured": true, 00:10:47.726 "data_offset": 2048, 00:10:47.726 "data_size": 63488 00:10:47.726 }, 00:10:47.726 { 00:10:47.726 "name": "BaseBdev4", 00:10:47.726 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:47.726 "is_configured": true, 00:10:47.726 "data_offset": 2048, 00:10:47.727 "data_size": 63488 00:10:47.727 } 00:10:47.727 ] 00:10:47.727 }' 00:10:47.727 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.727 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.040 [2024-10-29 10:59:53.453881] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:48.040 "name": "Existed_Raid", 00:10:48.040 "aliases": [ 00:10:48.040 "be6a985a-a380-4dc4-896f-af15c0a0b725" 00:10:48.040 ], 00:10:48.040 "product_name": "Raid Volume", 00:10:48.040 "block_size": 512, 00:10:48.040 "num_blocks": 253952, 00:10:48.040 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:48.040 "assigned_rate_limits": { 00:10:48.040 "rw_ios_per_sec": 0, 00:10:48.040 "rw_mbytes_per_sec": 0, 00:10:48.040 "r_mbytes_per_sec": 0, 00:10:48.040 "w_mbytes_per_sec": 0 00:10:48.040 }, 00:10:48.040 "claimed": false, 00:10:48.040 "zoned": false, 00:10:48.040 "supported_io_types": { 00:10:48.040 "read": true, 00:10:48.040 "write": true, 00:10:48.040 "unmap": true, 00:10:48.040 "flush": true, 00:10:48.040 "reset": true, 00:10:48.040 "nvme_admin": false, 00:10:48.040 "nvme_io": false, 00:10:48.040 "nvme_io_md": false, 00:10:48.040 "write_zeroes": true, 00:10:48.040 "zcopy": false, 00:10:48.040 "get_zone_info": false, 00:10:48.040 "zone_management": false, 00:10:48.040 "zone_append": false, 00:10:48.040 "compare": false, 00:10:48.040 "compare_and_write": false, 00:10:48.040 "abort": false, 00:10:48.040 "seek_hole": false, 00:10:48.040 "seek_data": false, 00:10:48.040 "copy": false, 00:10:48.040 "nvme_iov_md": false 00:10:48.040 }, 00:10:48.040 "memory_domains": [ 00:10:48.040 { 00:10:48.040 "dma_device_id": "system", 00:10:48.040 "dma_device_type": 1 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.040 "dma_device_type": 2 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "system", 00:10:48.040 "dma_device_type": 1 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.040 "dma_device_type": 2 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "system", 00:10:48.040 "dma_device_type": 1 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.040 "dma_device_type": 2 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "system", 00:10:48.040 "dma_device_type": 1 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.040 "dma_device_type": 2 00:10:48.040 } 00:10:48.040 ], 00:10:48.040 "driver_specific": { 00:10:48.040 "raid": { 00:10:48.040 "uuid": "be6a985a-a380-4dc4-896f-af15c0a0b725", 00:10:48.040 "strip_size_kb": 64, 00:10:48.040 "state": "online", 00:10:48.040 "raid_level": "concat", 00:10:48.040 "superblock": true, 00:10:48.040 "num_base_bdevs": 4, 00:10:48.040 "num_base_bdevs_discovered": 4, 00:10:48.040 "num_base_bdevs_operational": 4, 00:10:48.040 "base_bdevs_list": [ 00:10:48.040 { 00:10:48.040 "name": "NewBaseBdev", 00:10:48.040 "uuid": "c35b3820-177a-4e5a-89d8-813b455d6e2a", 00:10:48.040 "is_configured": true, 00:10:48.040 "data_offset": 2048, 00:10:48.040 "data_size": 63488 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "name": "BaseBdev2", 00:10:48.040 "uuid": "2d677b77-afe1-46a3-8141-46bf8ce017a1", 00:10:48.040 "is_configured": true, 00:10:48.040 "data_offset": 2048, 00:10:48.040 "data_size": 63488 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "name": "BaseBdev3", 00:10:48.040 "uuid": "2db5e54c-06e1-4c35-a929-b2206a81b5df", 00:10:48.040 "is_configured": true, 00:10:48.040 "data_offset": 2048, 00:10:48.040 "data_size": 63488 00:10:48.040 }, 00:10:48.040 { 00:10:48.040 "name": "BaseBdev4", 00:10:48.040 "uuid": "2dd17ce8-a75e-4f29-bb99-3ffe0d5d3ea3", 00:10:48.040 "is_configured": true, 00:10:48.040 "data_offset": 2048, 00:10:48.040 "data_size": 63488 00:10:48.040 } 00:10:48.040 ] 00:10:48.040 } 00:10:48.040 } 00:10:48.040 }' 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:48.040 BaseBdev2 00:10:48.040 BaseBdev3 00:10:48.040 BaseBdev4' 00:10:48.040 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.306 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.307 [2024-10-29 10:59:53.764885] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:48.307 [2024-10-29 10:59:53.764996] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:48.307 [2024-10-29 10:59:53.765110] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:48.307 [2024-10-29 10:59:53.765221] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:48.307 [2024-10-29 10:59:53.765277] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82977 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 82977 ']' 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 82977 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:48.307 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 82977 00:10:48.566 killing process with pid 82977 00:10:48.566 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:48.566 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:48.566 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 82977' 00:10:48.566 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 82977 00:10:48.566 [2024-10-29 10:59:53.812000] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:48.566 10:59:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 82977 00:10:48.566 [2024-10-29 10:59:53.888416] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:48.825 10:59:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:48.825 00:10:48.825 real 0m9.468s 00:10:48.825 user 0m15.850s 00:10:48.825 sys 0m2.016s 00:10:48.825 10:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:48.825 ************************************ 00:10:48.825 END TEST raid_state_function_test_sb 00:10:48.825 ************************************ 00:10:48.825 10:59:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.825 10:59:54 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:48.825 10:59:54 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:10:48.825 10:59:54 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:48.825 10:59:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:48.825 ************************************ 00:10:48.825 START TEST raid_superblock_test 00:10:48.825 ************************************ 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test concat 4 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:48.825 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83625 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83625 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 83625 ']' 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:48.825 10:59:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.084 [2024-10-29 10:59:54.380892] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:49.085 [2024-10-29 10:59:54.381106] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83625 ] 00:10:49.085 [2024-10-29 10:59:54.536997] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:49.085 [2024-10-29 10:59:54.577507] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:49.344 [2024-10-29 10:59:54.654072] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:49.344 [2024-10-29 10:59:54.654213] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 malloc1 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 [2024-10-29 10:59:55.233742] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:49.915 [2024-10-29 10:59:55.233895] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.915 [2024-10-29 10:59:55.233940] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:49.915 [2024-10-29 10:59:55.233979] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.915 [2024-10-29 10:59:55.236524] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.915 [2024-10-29 10:59:55.236567] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:49.915 pt1 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 malloc2 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 [2024-10-29 10:59:55.268682] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:49.915 [2024-10-29 10:59:55.268807] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.915 [2024-10-29 10:59:55.268869] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:49.915 [2024-10-29 10:59:55.268899] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.915 [2024-10-29 10:59:55.271560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.915 [2024-10-29 10:59:55.271636] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:49.915 pt2 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 malloc3 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.915 [2024-10-29 10:59:55.308155] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:49.915 [2024-10-29 10:59:55.308277] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.915 [2024-10-29 10:59:55.308317] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:49.915 [2024-10-29 10:59:55.308351] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.915 [2024-10-29 10:59:55.310902] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.915 [2024-10-29 10:59:55.310979] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:49.915 pt3 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.915 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.916 malloc4 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.916 [2024-10-29 10:59:55.356710] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:49.916 [2024-10-29 10:59:55.356829] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.916 [2024-10-29 10:59:55.356866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:49.916 [2024-10-29 10:59:55.356901] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.916 [2024-10-29 10:59:55.359483] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.916 [2024-10-29 10:59:55.359554] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:49.916 pt4 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.916 [2024-10-29 10:59:55.368780] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:49.916 [2024-10-29 10:59:55.371032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:49.916 [2024-10-29 10:59:55.371133] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:49.916 [2024-10-29 10:59:55.371217] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:49.916 [2024-10-29 10:59:55.371479] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:10:49.916 [2024-10-29 10:59:55.371532] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:49.916 [2024-10-29 10:59:55.371830] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:49.916 [2024-10-29 10:59:55.372058] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:10:49.916 [2024-10-29 10:59:55.372104] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:10:49.916 [2024-10-29 10:59:55.372320] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.916 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.176 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.176 "name": "raid_bdev1", 00:10:50.176 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:50.176 "strip_size_kb": 64, 00:10:50.176 "state": "online", 00:10:50.176 "raid_level": "concat", 00:10:50.176 "superblock": true, 00:10:50.176 "num_base_bdevs": 4, 00:10:50.176 "num_base_bdevs_discovered": 4, 00:10:50.176 "num_base_bdevs_operational": 4, 00:10:50.176 "base_bdevs_list": [ 00:10:50.176 { 00:10:50.176 "name": "pt1", 00:10:50.176 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:50.176 "is_configured": true, 00:10:50.176 "data_offset": 2048, 00:10:50.176 "data_size": 63488 00:10:50.176 }, 00:10:50.176 { 00:10:50.176 "name": "pt2", 00:10:50.176 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:50.176 "is_configured": true, 00:10:50.176 "data_offset": 2048, 00:10:50.176 "data_size": 63488 00:10:50.176 }, 00:10:50.176 { 00:10:50.176 "name": "pt3", 00:10:50.176 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:50.176 "is_configured": true, 00:10:50.176 "data_offset": 2048, 00:10:50.176 "data_size": 63488 00:10:50.176 }, 00:10:50.176 { 00:10:50.176 "name": "pt4", 00:10:50.176 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:50.176 "is_configured": true, 00:10:50.176 "data_offset": 2048, 00:10:50.176 "data_size": 63488 00:10:50.176 } 00:10:50.176 ] 00:10:50.176 }' 00:10:50.176 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.176 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:50.436 [2024-10-29 10:59:55.856261] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:50.436 "name": "raid_bdev1", 00:10:50.436 "aliases": [ 00:10:50.436 "f23aa60f-d875-4782-bec9-f9a6e815e2a8" 00:10:50.436 ], 00:10:50.436 "product_name": "Raid Volume", 00:10:50.436 "block_size": 512, 00:10:50.436 "num_blocks": 253952, 00:10:50.436 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:50.436 "assigned_rate_limits": { 00:10:50.436 "rw_ios_per_sec": 0, 00:10:50.436 "rw_mbytes_per_sec": 0, 00:10:50.436 "r_mbytes_per_sec": 0, 00:10:50.436 "w_mbytes_per_sec": 0 00:10:50.436 }, 00:10:50.436 "claimed": false, 00:10:50.436 "zoned": false, 00:10:50.436 "supported_io_types": { 00:10:50.436 "read": true, 00:10:50.436 "write": true, 00:10:50.436 "unmap": true, 00:10:50.436 "flush": true, 00:10:50.436 "reset": true, 00:10:50.436 "nvme_admin": false, 00:10:50.436 "nvme_io": false, 00:10:50.436 "nvme_io_md": false, 00:10:50.436 "write_zeroes": true, 00:10:50.436 "zcopy": false, 00:10:50.436 "get_zone_info": false, 00:10:50.436 "zone_management": false, 00:10:50.436 "zone_append": false, 00:10:50.436 "compare": false, 00:10:50.436 "compare_and_write": false, 00:10:50.436 "abort": false, 00:10:50.436 "seek_hole": false, 00:10:50.436 "seek_data": false, 00:10:50.436 "copy": false, 00:10:50.436 "nvme_iov_md": false 00:10:50.436 }, 00:10:50.436 "memory_domains": [ 00:10:50.436 { 00:10:50.436 "dma_device_id": "system", 00:10:50.436 "dma_device_type": 1 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.436 "dma_device_type": 2 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "system", 00:10:50.436 "dma_device_type": 1 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.436 "dma_device_type": 2 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "system", 00:10:50.436 "dma_device_type": 1 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.436 "dma_device_type": 2 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "system", 00:10:50.436 "dma_device_type": 1 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:50.436 "dma_device_type": 2 00:10:50.436 } 00:10:50.436 ], 00:10:50.436 "driver_specific": { 00:10:50.436 "raid": { 00:10:50.436 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:50.436 "strip_size_kb": 64, 00:10:50.436 "state": "online", 00:10:50.436 "raid_level": "concat", 00:10:50.436 "superblock": true, 00:10:50.436 "num_base_bdevs": 4, 00:10:50.436 "num_base_bdevs_discovered": 4, 00:10:50.436 "num_base_bdevs_operational": 4, 00:10:50.436 "base_bdevs_list": [ 00:10:50.436 { 00:10:50.436 "name": "pt1", 00:10:50.436 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:50.436 "is_configured": true, 00:10:50.436 "data_offset": 2048, 00:10:50.436 "data_size": 63488 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "name": "pt2", 00:10:50.436 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:50.436 "is_configured": true, 00:10:50.436 "data_offset": 2048, 00:10:50.436 "data_size": 63488 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "name": "pt3", 00:10:50.436 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:50.436 "is_configured": true, 00:10:50.436 "data_offset": 2048, 00:10:50.436 "data_size": 63488 00:10:50.436 }, 00:10:50.436 { 00:10:50.436 "name": "pt4", 00:10:50.436 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:50.436 "is_configured": true, 00:10:50.436 "data_offset": 2048, 00:10:50.436 "data_size": 63488 00:10:50.436 } 00:10:50.436 ] 00:10:50.436 } 00:10:50.436 } 00:10:50.436 }' 00:10:50.436 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:50.696 pt2 00:10:50.696 pt3 00:10:50.696 pt4' 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.696 10:59:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:50.696 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.697 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.697 [2024-10-29 10:59:56.183630] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f23aa60f-d875-4782-bec9-f9a6e815e2a8 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f23aa60f-d875-4782-bec9-f9a6e815e2a8 ']' 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 [2024-10-29 10:59:56.227222] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.957 [2024-10-29 10:59:56.227311] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:50.957 [2024-10-29 10:59:56.227471] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:50.957 [2024-10-29 10:59:56.227616] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:50.957 [2024-10-29 10:59:56.227676] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.957 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.957 [2024-10-29 10:59:56.391015] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:50.958 [2024-10-29 10:59:56.393476] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:50.958 [2024-10-29 10:59:56.393578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:50.958 [2024-10-29 10:59:56.393630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:50.958 [2024-10-29 10:59:56.393716] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:50.958 [2024-10-29 10:59:56.393834] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:50.958 [2024-10-29 10:59:56.393911] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:50.958 [2024-10-29 10:59:56.394006] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:50.958 [2024-10-29 10:59:56.394064] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.958 [2024-10-29 10:59:56.394105] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:10:50.958 request: 00:10:50.958 { 00:10:50.958 "name": "raid_bdev1", 00:10:50.958 "raid_level": "concat", 00:10:50.958 "base_bdevs": [ 00:10:50.958 "malloc1", 00:10:50.958 "malloc2", 00:10:50.958 "malloc3", 00:10:50.958 "malloc4" 00:10:50.958 ], 00:10:50.958 "strip_size_kb": 64, 00:10:50.958 "superblock": false, 00:10:50.958 "method": "bdev_raid_create", 00:10:50.958 "req_id": 1 00:10:50.958 } 00:10:50.958 Got JSON-RPC error response 00:10:50.958 response: 00:10:50.958 { 00:10:50.958 "code": -17, 00:10:50.958 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:50.958 } 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.958 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.218 [2024-10-29 10:59:56.458819] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:51.218 [2024-10-29 10:59:56.458928] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.218 [2024-10-29 10:59:56.458971] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:51.218 [2024-10-29 10:59:56.458999] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.218 [2024-10-29 10:59:56.461689] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.218 [2024-10-29 10:59:56.461758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:51.218 [2024-10-29 10:59:56.461866] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:51.218 [2024-10-29 10:59:56.461940] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:51.218 pt1 00:10:51.218 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.218 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:51.218 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:51.218 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.219 "name": "raid_bdev1", 00:10:51.219 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:51.219 "strip_size_kb": 64, 00:10:51.219 "state": "configuring", 00:10:51.219 "raid_level": "concat", 00:10:51.219 "superblock": true, 00:10:51.219 "num_base_bdevs": 4, 00:10:51.219 "num_base_bdevs_discovered": 1, 00:10:51.219 "num_base_bdevs_operational": 4, 00:10:51.219 "base_bdevs_list": [ 00:10:51.219 { 00:10:51.219 "name": "pt1", 00:10:51.219 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:51.219 "is_configured": true, 00:10:51.219 "data_offset": 2048, 00:10:51.219 "data_size": 63488 00:10:51.219 }, 00:10:51.219 { 00:10:51.219 "name": null, 00:10:51.219 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:51.219 "is_configured": false, 00:10:51.219 "data_offset": 2048, 00:10:51.219 "data_size": 63488 00:10:51.219 }, 00:10:51.219 { 00:10:51.219 "name": null, 00:10:51.219 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:51.219 "is_configured": false, 00:10:51.219 "data_offset": 2048, 00:10:51.219 "data_size": 63488 00:10:51.219 }, 00:10:51.219 { 00:10:51.219 "name": null, 00:10:51.219 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:51.219 "is_configured": false, 00:10:51.219 "data_offset": 2048, 00:10:51.219 "data_size": 63488 00:10:51.219 } 00:10:51.219 ] 00:10:51.219 }' 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.219 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.479 [2024-10-29 10:59:56.898129] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:51.479 [2024-10-29 10:59:56.898281] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.479 [2024-10-29 10:59:56.898323] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:51.479 [2024-10-29 10:59:56.898353] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.479 [2024-10-29 10:59:56.898930] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.479 [2024-10-29 10:59:56.898990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:51.479 [2024-10-29 10:59:56.899137] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:51.479 [2024-10-29 10:59:56.899192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:51.479 pt2 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.479 [2024-10-29 10:59:56.910075] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.479 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.479 "name": "raid_bdev1", 00:10:51.479 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:51.479 "strip_size_kb": 64, 00:10:51.479 "state": "configuring", 00:10:51.479 "raid_level": "concat", 00:10:51.479 "superblock": true, 00:10:51.479 "num_base_bdevs": 4, 00:10:51.479 "num_base_bdevs_discovered": 1, 00:10:51.479 "num_base_bdevs_operational": 4, 00:10:51.479 "base_bdevs_list": [ 00:10:51.479 { 00:10:51.479 "name": "pt1", 00:10:51.479 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:51.479 "is_configured": true, 00:10:51.479 "data_offset": 2048, 00:10:51.479 "data_size": 63488 00:10:51.479 }, 00:10:51.479 { 00:10:51.479 "name": null, 00:10:51.479 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:51.479 "is_configured": false, 00:10:51.479 "data_offset": 0, 00:10:51.479 "data_size": 63488 00:10:51.479 }, 00:10:51.479 { 00:10:51.479 "name": null, 00:10:51.479 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:51.479 "is_configured": false, 00:10:51.479 "data_offset": 2048, 00:10:51.479 "data_size": 63488 00:10:51.479 }, 00:10:51.479 { 00:10:51.479 "name": null, 00:10:51.479 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:51.479 "is_configured": false, 00:10:51.479 "data_offset": 2048, 00:10:51.480 "data_size": 63488 00:10:51.480 } 00:10:51.480 ] 00:10:51.480 }' 00:10:51.480 10:59:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.480 10:59:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.050 [2024-10-29 10:59:57.321407] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:52.050 [2024-10-29 10:59:57.321498] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:52.050 [2024-10-29 10:59:57.321519] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:10:52.050 [2024-10-29 10:59:57.321532] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:52.050 [2024-10-29 10:59:57.322024] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:52.050 [2024-10-29 10:59:57.322045] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:52.050 [2024-10-29 10:59:57.322135] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:52.050 [2024-10-29 10:59:57.322164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:52.050 pt2 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.050 [2024-10-29 10:59:57.333335] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:52.050 [2024-10-29 10:59:57.333481] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:52.050 [2024-10-29 10:59:57.333518] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:52.050 [2024-10-29 10:59:57.333565] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:52.050 [2024-10-29 10:59:57.334007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:52.050 [2024-10-29 10:59:57.334083] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:52.050 [2024-10-29 10:59:57.334189] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:52.050 [2024-10-29 10:59:57.334248] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:52.050 pt3 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.050 [2024-10-29 10:59:57.345268] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:52.050 [2024-10-29 10:59:57.345377] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:52.050 [2024-10-29 10:59:57.345414] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:52.050 [2024-10-29 10:59:57.345450] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:52.050 [2024-10-29 10:59:57.345856] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:52.050 [2024-10-29 10:59:57.345915] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:52.050 [2024-10-29 10:59:57.346025] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:52.050 [2024-10-29 10:59:57.346080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:52.050 [2024-10-29 10:59:57.346243] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:10:52.050 [2024-10-29 10:59:57.346284] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:52.050 [2024-10-29 10:59:57.346576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:10:52.050 [2024-10-29 10:59:57.346757] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:10:52.050 [2024-10-29 10:59:57.346772] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:10:52.050 [2024-10-29 10:59:57.346889] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:52.050 pt4 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.050 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:52.050 "name": "raid_bdev1", 00:10:52.050 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:52.050 "strip_size_kb": 64, 00:10:52.050 "state": "online", 00:10:52.050 "raid_level": "concat", 00:10:52.050 "superblock": true, 00:10:52.050 "num_base_bdevs": 4, 00:10:52.050 "num_base_bdevs_discovered": 4, 00:10:52.050 "num_base_bdevs_operational": 4, 00:10:52.050 "base_bdevs_list": [ 00:10:52.050 { 00:10:52.050 "name": "pt1", 00:10:52.050 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:52.050 "is_configured": true, 00:10:52.050 "data_offset": 2048, 00:10:52.050 "data_size": 63488 00:10:52.051 }, 00:10:52.051 { 00:10:52.051 "name": "pt2", 00:10:52.051 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:52.051 "is_configured": true, 00:10:52.051 "data_offset": 2048, 00:10:52.051 "data_size": 63488 00:10:52.051 }, 00:10:52.051 { 00:10:52.051 "name": "pt3", 00:10:52.051 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:52.051 "is_configured": true, 00:10:52.051 "data_offset": 2048, 00:10:52.051 "data_size": 63488 00:10:52.051 }, 00:10:52.051 { 00:10:52.051 "name": "pt4", 00:10:52.051 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:52.051 "is_configured": true, 00:10:52.051 "data_offset": 2048, 00:10:52.051 "data_size": 63488 00:10:52.051 } 00:10:52.051 ] 00:10:52.051 }' 00:10:52.051 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:52.051 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.311 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.311 [2024-10-29 10:59:57.796918] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:52.571 "name": "raid_bdev1", 00:10:52.571 "aliases": [ 00:10:52.571 "f23aa60f-d875-4782-bec9-f9a6e815e2a8" 00:10:52.571 ], 00:10:52.571 "product_name": "Raid Volume", 00:10:52.571 "block_size": 512, 00:10:52.571 "num_blocks": 253952, 00:10:52.571 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:52.571 "assigned_rate_limits": { 00:10:52.571 "rw_ios_per_sec": 0, 00:10:52.571 "rw_mbytes_per_sec": 0, 00:10:52.571 "r_mbytes_per_sec": 0, 00:10:52.571 "w_mbytes_per_sec": 0 00:10:52.571 }, 00:10:52.571 "claimed": false, 00:10:52.571 "zoned": false, 00:10:52.571 "supported_io_types": { 00:10:52.571 "read": true, 00:10:52.571 "write": true, 00:10:52.571 "unmap": true, 00:10:52.571 "flush": true, 00:10:52.571 "reset": true, 00:10:52.571 "nvme_admin": false, 00:10:52.571 "nvme_io": false, 00:10:52.571 "nvme_io_md": false, 00:10:52.571 "write_zeroes": true, 00:10:52.571 "zcopy": false, 00:10:52.571 "get_zone_info": false, 00:10:52.571 "zone_management": false, 00:10:52.571 "zone_append": false, 00:10:52.571 "compare": false, 00:10:52.571 "compare_and_write": false, 00:10:52.571 "abort": false, 00:10:52.571 "seek_hole": false, 00:10:52.571 "seek_data": false, 00:10:52.571 "copy": false, 00:10:52.571 "nvme_iov_md": false 00:10:52.571 }, 00:10:52.571 "memory_domains": [ 00:10:52.571 { 00:10:52.571 "dma_device_id": "system", 00:10:52.571 "dma_device_type": 1 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.571 "dma_device_type": 2 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "system", 00:10:52.571 "dma_device_type": 1 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.571 "dma_device_type": 2 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "system", 00:10:52.571 "dma_device_type": 1 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.571 "dma_device_type": 2 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "system", 00:10:52.571 "dma_device_type": 1 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:52.571 "dma_device_type": 2 00:10:52.571 } 00:10:52.571 ], 00:10:52.571 "driver_specific": { 00:10:52.571 "raid": { 00:10:52.571 "uuid": "f23aa60f-d875-4782-bec9-f9a6e815e2a8", 00:10:52.571 "strip_size_kb": 64, 00:10:52.571 "state": "online", 00:10:52.571 "raid_level": "concat", 00:10:52.571 "superblock": true, 00:10:52.571 "num_base_bdevs": 4, 00:10:52.571 "num_base_bdevs_discovered": 4, 00:10:52.571 "num_base_bdevs_operational": 4, 00:10:52.571 "base_bdevs_list": [ 00:10:52.571 { 00:10:52.571 "name": "pt1", 00:10:52.571 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:52.571 "is_configured": true, 00:10:52.571 "data_offset": 2048, 00:10:52.571 "data_size": 63488 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "name": "pt2", 00:10:52.571 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:52.571 "is_configured": true, 00:10:52.571 "data_offset": 2048, 00:10:52.571 "data_size": 63488 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "name": "pt3", 00:10:52.571 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:52.571 "is_configured": true, 00:10:52.571 "data_offset": 2048, 00:10:52.571 "data_size": 63488 00:10:52.571 }, 00:10:52.571 { 00:10:52.571 "name": "pt4", 00:10:52.571 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:52.571 "is_configured": true, 00:10:52.571 "data_offset": 2048, 00:10:52.571 "data_size": 63488 00:10:52.571 } 00:10:52.571 ] 00:10:52.571 } 00:10:52.571 } 00:10:52.571 }' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:52.571 pt2 00:10:52.571 pt3 00:10:52.571 pt4' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.571 10:59:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:52.571 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:52.572 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.572 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.572 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:52.832 [2024-10-29 10:59:58.100323] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f23aa60f-d875-4782-bec9-f9a6e815e2a8 '!=' f23aa60f-d875-4782-bec9-f9a6e815e2a8 ']' 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83625 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 83625 ']' 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 83625 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 83625 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 83625' 00:10:52.832 killing process with pid 83625 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 83625 00:10:52.832 [2024-10-29 10:59:58.180028] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:52.832 [2024-10-29 10:59:58.180153] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:52.832 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 83625 00:10:52.832 [2024-10-29 10:59:58.180238] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:52.832 [2024-10-29 10:59:58.180254] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:10:52.832 [2024-10-29 10:59:58.260044] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:53.092 ************************************ 00:10:53.092 END TEST raid_superblock_test 00:10:53.092 ************************************ 00:10:53.092 10:59:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:53.092 00:10:53.092 real 0m4.294s 00:10:53.092 user 0m6.563s 00:10:53.092 sys 0m0.967s 00:10:53.092 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:53.092 10:59:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.352 10:59:58 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:53.352 10:59:58 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:53.352 10:59:58 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:53.352 10:59:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:53.352 ************************************ 00:10:53.352 START TEST raid_read_error_test 00:10:53.352 ************************************ 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 4 read 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.GZ9xx03Njt 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83879 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83879 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 83879 ']' 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:53.352 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:53.352 10:59:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.352 [2024-10-29 10:59:58.762817] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:53.352 [2024-10-29 10:59:58.762942] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83879 ] 00:10:53.612 [2024-10-29 10:59:58.937313] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:53.612 [2024-10-29 10:59:58.976647] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:53.612 [2024-10-29 10:59:59.053179] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:53.612 [2024-10-29 10:59:59.053221] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.182 BaseBdev1_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.182 true 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.182 [2024-10-29 10:59:59.639982] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:54.182 [2024-10-29 10:59:59.640049] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.182 [2024-10-29 10:59:59.640073] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:54.182 [2024-10-29 10:59:59.640084] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.182 [2024-10-29 10:59:59.642615] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.182 [2024-10-29 10:59:59.642649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:54.182 BaseBdev1 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.182 BaseBdev2_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.182 true 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.182 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.441 [2024-10-29 10:59:59.682924] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:54.441 [2024-10-29 10:59:59.682994] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.441 [2024-10-29 10:59:59.683015] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:54.441 [2024-10-29 10:59:59.683024] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.441 [2024-10-29 10:59:59.685500] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.441 [2024-10-29 10:59:59.685534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:54.441 BaseBdev2 00:10:54.441 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.441 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:54.441 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:54.441 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.441 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.441 BaseBdev3_malloc 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 true 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 [2024-10-29 10:59:59.730096] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:54.442 [2024-10-29 10:59:59.730153] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.442 [2024-10-29 10:59:59.730175] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:54.442 [2024-10-29 10:59:59.730185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.442 [2024-10-29 10:59:59.732578] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.442 [2024-10-29 10:59:59.732623] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:54.442 BaseBdev3 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 BaseBdev4_malloc 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 true 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 [2024-10-29 10:59:59.787194] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:54.442 [2024-10-29 10:59:59.787247] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.442 [2024-10-29 10:59:59.787271] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:54.442 [2024-10-29 10:59:59.787279] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.442 [2024-10-29 10:59:59.789616] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.442 [2024-10-29 10:59:59.789649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:54.442 BaseBdev4 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 [2024-10-29 10:59:59.799255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:54.442 [2024-10-29 10:59:59.801361] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:54.442 [2024-10-29 10:59:59.801454] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:54.442 [2024-10-29 10:59:59.801504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:54.442 [2024-10-29 10:59:59.801694] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:54.442 [2024-10-29 10:59:59.801711] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:54.442 [2024-10-29 10:59:59.801988] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:54.442 [2024-10-29 10:59:59.802144] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:54.442 [2024-10-29 10:59:59.802178] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:54.442 [2024-10-29 10:59:59.802308] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:54.442 "name": "raid_bdev1", 00:10:54.442 "uuid": "beb7c5cd-0788-4013-9b36-bd5efa5c96f4", 00:10:54.442 "strip_size_kb": 64, 00:10:54.442 "state": "online", 00:10:54.442 "raid_level": "concat", 00:10:54.442 "superblock": true, 00:10:54.442 "num_base_bdevs": 4, 00:10:54.442 "num_base_bdevs_discovered": 4, 00:10:54.442 "num_base_bdevs_operational": 4, 00:10:54.442 "base_bdevs_list": [ 00:10:54.442 { 00:10:54.442 "name": "BaseBdev1", 00:10:54.442 "uuid": "9ff686cf-4035-59c1-a18b-d4f87c808a86", 00:10:54.442 "is_configured": true, 00:10:54.442 "data_offset": 2048, 00:10:54.442 "data_size": 63488 00:10:54.442 }, 00:10:54.442 { 00:10:54.442 "name": "BaseBdev2", 00:10:54.442 "uuid": "12b54cc6-1755-5df3-8afc-132604a427c6", 00:10:54.442 "is_configured": true, 00:10:54.442 "data_offset": 2048, 00:10:54.442 "data_size": 63488 00:10:54.442 }, 00:10:54.442 { 00:10:54.442 "name": "BaseBdev3", 00:10:54.442 "uuid": "36fa648c-c275-5f3c-9fa8-4a01cc3dd44f", 00:10:54.442 "is_configured": true, 00:10:54.442 "data_offset": 2048, 00:10:54.442 "data_size": 63488 00:10:54.442 }, 00:10:54.442 { 00:10:54.442 "name": "BaseBdev4", 00:10:54.442 "uuid": "c2108f64-6837-577d-a9d7-2a854e0c85fb", 00:10:54.442 "is_configured": true, 00:10:54.442 "data_offset": 2048, 00:10:54.442 "data_size": 63488 00:10:54.442 } 00:10:54.442 ] 00:10:54.442 }' 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:54.442 10:59:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.010 11:00:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:55.010 11:00:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:55.010 [2024-10-29 11:00:00.310873] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.948 "name": "raid_bdev1", 00:10:55.948 "uuid": "beb7c5cd-0788-4013-9b36-bd5efa5c96f4", 00:10:55.948 "strip_size_kb": 64, 00:10:55.948 "state": "online", 00:10:55.948 "raid_level": "concat", 00:10:55.948 "superblock": true, 00:10:55.948 "num_base_bdevs": 4, 00:10:55.948 "num_base_bdevs_discovered": 4, 00:10:55.948 "num_base_bdevs_operational": 4, 00:10:55.948 "base_bdevs_list": [ 00:10:55.948 { 00:10:55.948 "name": "BaseBdev1", 00:10:55.948 "uuid": "9ff686cf-4035-59c1-a18b-d4f87c808a86", 00:10:55.948 "is_configured": true, 00:10:55.948 "data_offset": 2048, 00:10:55.948 "data_size": 63488 00:10:55.948 }, 00:10:55.948 { 00:10:55.948 "name": "BaseBdev2", 00:10:55.948 "uuid": "12b54cc6-1755-5df3-8afc-132604a427c6", 00:10:55.948 "is_configured": true, 00:10:55.948 "data_offset": 2048, 00:10:55.948 "data_size": 63488 00:10:55.948 }, 00:10:55.948 { 00:10:55.948 "name": "BaseBdev3", 00:10:55.948 "uuid": "36fa648c-c275-5f3c-9fa8-4a01cc3dd44f", 00:10:55.948 "is_configured": true, 00:10:55.948 "data_offset": 2048, 00:10:55.948 "data_size": 63488 00:10:55.948 }, 00:10:55.948 { 00:10:55.948 "name": "BaseBdev4", 00:10:55.948 "uuid": "c2108f64-6837-577d-a9d7-2a854e0c85fb", 00:10:55.948 "is_configured": true, 00:10:55.948 "data_offset": 2048, 00:10:55.948 "data_size": 63488 00:10:55.948 } 00:10:55.948 ] 00:10:55.948 }' 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.948 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.206 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:56.206 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:56.206 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.206 [2024-10-29 11:00:01.675468] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:56.206 [2024-10-29 11:00:01.675516] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:56.206 [2024-10-29 11:00:01.678050] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:56.206 [2024-10-29 11:00:01.678111] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:56.206 [2024-10-29 11:00:01.678164] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:56.206 [2024-10-29 11:00:01.678175] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:56.206 { 00:10:56.206 "results": [ 00:10:56.206 { 00:10:56.206 "job": "raid_bdev1", 00:10:56.206 "core_mask": "0x1", 00:10:56.206 "workload": "randrw", 00:10:56.206 "percentage": 50, 00:10:56.206 "status": "finished", 00:10:56.206 "queue_depth": 1, 00:10:56.206 "io_size": 131072, 00:10:56.206 "runtime": 1.365093, 00:10:56.206 "iops": 14377.042443262108, 00:10:56.206 "mibps": 1797.1303054077634, 00:10:56.206 "io_failed": 1, 00:10:56.206 "io_timeout": 0, 00:10:56.206 "avg_latency_us": 97.95988175098779, 00:10:56.206 "min_latency_us": 24.370305676855896, 00:10:56.206 "max_latency_us": 1373.6803493449781 00:10:56.206 } 00:10:56.206 ], 00:10:56.206 "core_count": 1 00:10:56.206 } 00:10:56.206 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83879 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 83879 ']' 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 83879 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:56.207 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 83879 00:10:56.466 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:56.466 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:56.466 killing process with pid 83879 00:10:56.466 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 83879' 00:10:56.466 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 83879 00:10:56.466 [2024-10-29 11:00:01.714479] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:56.466 11:00:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 83879 00:10:56.466 [2024-10-29 11:00:01.784650] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.GZ9xx03Njt 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:56.726 00:10:56.726 real 0m3.459s 00:10:56.726 user 0m4.219s 00:10:56.726 sys 0m0.613s 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:10:56.726 11:00:02 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.726 ************************************ 00:10:56.726 END TEST raid_read_error_test 00:10:56.726 ************************************ 00:10:56.726 11:00:02 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:56.726 11:00:02 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:10:56.726 11:00:02 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:10:56.726 11:00:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:56.726 ************************************ 00:10:56.726 START TEST raid_write_error_test 00:10:56.726 ************************************ 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test concat 4 write 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.suABCQ5tqE 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=84008 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 84008 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 84008 ']' 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:10:56.726 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:10:56.726 11:00:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.987 [2024-10-29 11:00:02.297111] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:10:56.987 [2024-10-29 11:00:02.297240] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84008 ] 00:10:56.987 [2024-10-29 11:00:02.467077] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:57.247 [2024-10-29 11:00:02.508359] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:57.247 [2024-10-29 11:00:02.585132] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:57.247 [2024-10-29 11:00:02.585177] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 BaseBdev1_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 true 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 [2024-10-29 11:00:03.131377] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:57.817 [2024-10-29 11:00:03.131463] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.817 [2024-10-29 11:00:03.131491] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:10:57.817 [2024-10-29 11:00:03.131500] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.817 [2024-10-29 11:00:03.133891] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.817 [2024-10-29 11:00:03.133925] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:57.817 BaseBdev1 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 BaseBdev2_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 true 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 [2024-10-29 11:00:03.177924] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:57.817 [2024-10-29 11:00:03.177978] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.817 [2024-10-29 11:00:03.177996] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:57.817 [2024-10-29 11:00:03.178006] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.817 [2024-10-29 11:00:03.180369] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.817 [2024-10-29 11:00:03.180398] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:57.817 BaseBdev2 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 BaseBdev3_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 true 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.817 [2024-10-29 11:00:03.224730] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:57.817 [2024-10-29 11:00:03.224781] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.817 [2024-10-29 11:00:03.224801] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:10:57.817 [2024-10-29 11:00:03.224810] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.817 [2024-10-29 11:00:03.227116] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.817 [2024-10-29 11:00:03.227150] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:57.817 BaseBdev3 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.817 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.818 BaseBdev4_malloc 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.818 true 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.818 [2024-10-29 11:00:03.281022] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:57.818 [2024-10-29 11:00:03.281074] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.818 [2024-10-29 11:00:03.281098] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:57.818 [2024-10-29 11:00:03.281106] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.818 [2024-10-29 11:00:03.283340] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.818 [2024-10-29 11:00:03.283390] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:57.818 BaseBdev4 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.818 [2024-10-29 11:00:03.293078] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:57.818 [2024-10-29 11:00:03.295116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:57.818 [2024-10-29 11:00:03.295199] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:57.818 [2024-10-29 11:00:03.295247] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:57.818 [2024-10-29 11:00:03.295482] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:10:57.818 [2024-10-29 11:00:03.295513] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:57.818 [2024-10-29 11:00:03.295795] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:10:57.818 [2024-10-29 11:00:03.295939] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:10:57.818 [2024-10-29 11:00:03.295959] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:10:57.818 [2024-10-29 11:00:03.296108] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.818 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.077 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.077 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.077 "name": "raid_bdev1", 00:10:58.077 "uuid": "b6a47116-25cf-4636-8cca-85eedf34e2f9", 00:10:58.077 "strip_size_kb": 64, 00:10:58.077 "state": "online", 00:10:58.077 "raid_level": "concat", 00:10:58.077 "superblock": true, 00:10:58.077 "num_base_bdevs": 4, 00:10:58.077 "num_base_bdevs_discovered": 4, 00:10:58.077 "num_base_bdevs_operational": 4, 00:10:58.077 "base_bdevs_list": [ 00:10:58.077 { 00:10:58.077 "name": "BaseBdev1", 00:10:58.077 "uuid": "bc205aa0-3ce3-57d6-885c-256c3112705c", 00:10:58.077 "is_configured": true, 00:10:58.077 "data_offset": 2048, 00:10:58.077 "data_size": 63488 00:10:58.077 }, 00:10:58.077 { 00:10:58.077 "name": "BaseBdev2", 00:10:58.077 "uuid": "56685150-7305-5c4a-bc92-c764f9a00577", 00:10:58.077 "is_configured": true, 00:10:58.077 "data_offset": 2048, 00:10:58.077 "data_size": 63488 00:10:58.077 }, 00:10:58.077 { 00:10:58.077 "name": "BaseBdev3", 00:10:58.078 "uuid": "a6e51598-a8a8-555b-98c4-85ea7a4460ce", 00:10:58.078 "is_configured": true, 00:10:58.078 "data_offset": 2048, 00:10:58.078 "data_size": 63488 00:10:58.078 }, 00:10:58.078 { 00:10:58.078 "name": "BaseBdev4", 00:10:58.078 "uuid": "cb30f0f8-d4a2-597d-b294-e2eae759a2f8", 00:10:58.078 "is_configured": true, 00:10:58.078 "data_offset": 2048, 00:10:58.078 "data_size": 63488 00:10:58.078 } 00:10:58.078 ] 00:10:58.078 }' 00:10:58.078 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.078 11:00:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.337 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:58.337 11:00:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:58.337 [2024-10-29 11:00:03.828674] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.278 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.539 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:59.539 "name": "raid_bdev1", 00:10:59.539 "uuid": "b6a47116-25cf-4636-8cca-85eedf34e2f9", 00:10:59.539 "strip_size_kb": 64, 00:10:59.539 "state": "online", 00:10:59.539 "raid_level": "concat", 00:10:59.539 "superblock": true, 00:10:59.539 "num_base_bdevs": 4, 00:10:59.539 "num_base_bdevs_discovered": 4, 00:10:59.539 "num_base_bdevs_operational": 4, 00:10:59.539 "base_bdevs_list": [ 00:10:59.539 { 00:10:59.539 "name": "BaseBdev1", 00:10:59.539 "uuid": "bc205aa0-3ce3-57d6-885c-256c3112705c", 00:10:59.539 "is_configured": true, 00:10:59.539 "data_offset": 2048, 00:10:59.539 "data_size": 63488 00:10:59.539 }, 00:10:59.539 { 00:10:59.539 "name": "BaseBdev2", 00:10:59.539 "uuid": "56685150-7305-5c4a-bc92-c764f9a00577", 00:10:59.539 "is_configured": true, 00:10:59.539 "data_offset": 2048, 00:10:59.539 "data_size": 63488 00:10:59.539 }, 00:10:59.539 { 00:10:59.539 "name": "BaseBdev3", 00:10:59.539 "uuid": "a6e51598-a8a8-555b-98c4-85ea7a4460ce", 00:10:59.539 "is_configured": true, 00:10:59.539 "data_offset": 2048, 00:10:59.539 "data_size": 63488 00:10:59.539 }, 00:10:59.539 { 00:10:59.539 "name": "BaseBdev4", 00:10:59.539 "uuid": "cb30f0f8-d4a2-597d-b294-e2eae759a2f8", 00:10:59.539 "is_configured": true, 00:10:59.539 "data_offset": 2048, 00:10:59.539 "data_size": 63488 00:10:59.539 } 00:10:59.539 ] 00:10:59.539 }' 00:10:59.539 11:00:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:59.539 11:00:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.799 [2024-10-29 11:00:05.165157] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:59.799 [2024-10-29 11:00:05.165207] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:59.799 [2024-10-29 11:00:05.167822] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:59.799 [2024-10-29 11:00:05.167901] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:59.799 [2024-10-29 11:00:05.167962] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:59.799 [2024-10-29 11:00:05.167972] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:10:59.799 { 00:10:59.799 "results": [ 00:10:59.799 { 00:10:59.799 "job": "raid_bdev1", 00:10:59.799 "core_mask": "0x1", 00:10:59.799 "workload": "randrw", 00:10:59.799 "percentage": 50, 00:10:59.799 "status": "finished", 00:10:59.799 "queue_depth": 1, 00:10:59.799 "io_size": 131072, 00:10:59.799 "runtime": 1.336973, 00:10:59.799 "iops": 14297.970116075643, 00:10:59.799 "mibps": 1787.2462645094554, 00:10:59.799 "io_failed": 1, 00:10:59.799 "io_timeout": 0, 00:10:59.799 "avg_latency_us": 98.40974317424329, 00:10:59.799 "min_latency_us": 24.705676855895195, 00:10:59.799 "max_latency_us": 1352.216593886463 00:10:59.799 } 00:10:59.799 ], 00:10:59.799 "core_count": 1 00:10:59.799 } 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 84008 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 84008 ']' 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 84008 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 84008 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:10:59.799 killing process with pid 84008 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 84008' 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 84008 00:10:59.799 [2024-10-29 11:00:05.212954] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:59.799 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 84008 00:10:59.799 [2024-10-29 11:00:05.278701] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:00.369 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.suABCQ5tqE 00:11:00.369 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:00.369 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:11:00.370 00:11:00.370 real 0m3.424s 00:11:00.370 user 0m4.134s 00:11:00.370 sys 0m0.633s 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:00.370 11:00:05 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.370 ************************************ 00:11:00.370 END TEST raid_write_error_test 00:11:00.370 ************************************ 00:11:00.370 11:00:05 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:11:00.370 11:00:05 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:11:00.370 11:00:05 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:11:00.370 11:00:05 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:00.370 11:00:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:00.370 ************************************ 00:11:00.370 START TEST raid_state_function_test 00:11:00.370 ************************************ 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 4 false 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=84135 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84135' 00:11:00.370 Process raid pid: 84135 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 84135 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 84135 ']' 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:00.370 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:00.370 11:00:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.370 [2024-10-29 11:00:05.775417] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:00.370 [2024-10-29 11:00:05.775550] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:00.629 [2024-10-29 11:00:05.946524] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:00.629 [2024-10-29 11:00:05.985299] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:00.629 [2024-10-29 11:00:06.063713] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.629 [2024-10-29 11:00:06.063753] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.202 [2024-10-29 11:00:06.616674] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:01.202 [2024-10-29 11:00:06.616742] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:01.202 [2024-10-29 11:00:06.616759] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:01.202 [2024-10-29 11:00:06.616771] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:01.202 [2024-10-29 11:00:06.616778] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:01.202 [2024-10-29 11:00:06.616790] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:01.202 [2024-10-29 11:00:06.616803] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:01.202 [2024-10-29 11:00:06.616813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.202 "name": "Existed_Raid", 00:11:01.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.202 "strip_size_kb": 0, 00:11:01.202 "state": "configuring", 00:11:01.202 "raid_level": "raid1", 00:11:01.202 "superblock": false, 00:11:01.202 "num_base_bdevs": 4, 00:11:01.202 "num_base_bdevs_discovered": 0, 00:11:01.202 "num_base_bdevs_operational": 4, 00:11:01.202 "base_bdevs_list": [ 00:11:01.202 { 00:11:01.202 "name": "BaseBdev1", 00:11:01.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.202 "is_configured": false, 00:11:01.202 "data_offset": 0, 00:11:01.202 "data_size": 0 00:11:01.202 }, 00:11:01.202 { 00:11:01.202 "name": "BaseBdev2", 00:11:01.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.202 "is_configured": false, 00:11:01.202 "data_offset": 0, 00:11:01.202 "data_size": 0 00:11:01.202 }, 00:11:01.202 { 00:11:01.202 "name": "BaseBdev3", 00:11:01.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.202 "is_configured": false, 00:11:01.202 "data_offset": 0, 00:11:01.202 "data_size": 0 00:11:01.202 }, 00:11:01.202 { 00:11:01.202 "name": "BaseBdev4", 00:11:01.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.202 "is_configured": false, 00:11:01.202 "data_offset": 0, 00:11:01.202 "data_size": 0 00:11:01.202 } 00:11:01.202 ] 00:11:01.202 }' 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.202 11:00:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.786 [2024-10-29 11:00:07.083768] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:01.786 [2024-10-29 11:00:07.083833] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.786 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.786 [2024-10-29 11:00:07.091756] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:01.786 [2024-10-29 11:00:07.091802] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:01.786 [2024-10-29 11:00:07.091811] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:01.786 [2024-10-29 11:00:07.091821] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:01.786 [2024-10-29 11:00:07.091827] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:01.786 [2024-10-29 11:00:07.091837] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:01.786 [2024-10-29 11:00:07.091842] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:01.786 [2024-10-29 11:00:07.091852] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.787 [2024-10-29 11:00:07.115125] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:01.787 BaseBdev1 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.787 [ 00:11:01.787 { 00:11:01.787 "name": "BaseBdev1", 00:11:01.787 "aliases": [ 00:11:01.787 "83f8c037-5b86-4cc6-8155-4ee7e9ad5590" 00:11:01.787 ], 00:11:01.787 "product_name": "Malloc disk", 00:11:01.787 "block_size": 512, 00:11:01.787 "num_blocks": 65536, 00:11:01.787 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:01.787 "assigned_rate_limits": { 00:11:01.787 "rw_ios_per_sec": 0, 00:11:01.787 "rw_mbytes_per_sec": 0, 00:11:01.787 "r_mbytes_per_sec": 0, 00:11:01.787 "w_mbytes_per_sec": 0 00:11:01.787 }, 00:11:01.787 "claimed": true, 00:11:01.787 "claim_type": "exclusive_write", 00:11:01.787 "zoned": false, 00:11:01.787 "supported_io_types": { 00:11:01.787 "read": true, 00:11:01.787 "write": true, 00:11:01.787 "unmap": true, 00:11:01.787 "flush": true, 00:11:01.787 "reset": true, 00:11:01.787 "nvme_admin": false, 00:11:01.787 "nvme_io": false, 00:11:01.787 "nvme_io_md": false, 00:11:01.787 "write_zeroes": true, 00:11:01.787 "zcopy": true, 00:11:01.787 "get_zone_info": false, 00:11:01.787 "zone_management": false, 00:11:01.787 "zone_append": false, 00:11:01.787 "compare": false, 00:11:01.787 "compare_and_write": false, 00:11:01.787 "abort": true, 00:11:01.787 "seek_hole": false, 00:11:01.787 "seek_data": false, 00:11:01.787 "copy": true, 00:11:01.787 "nvme_iov_md": false 00:11:01.787 }, 00:11:01.787 "memory_domains": [ 00:11:01.787 { 00:11:01.787 "dma_device_id": "system", 00:11:01.787 "dma_device_type": 1 00:11:01.787 }, 00:11:01.787 { 00:11:01.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.787 "dma_device_type": 2 00:11:01.787 } 00:11:01.787 ], 00:11:01.787 "driver_specific": {} 00:11:01.787 } 00:11:01.787 ] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.787 "name": "Existed_Raid", 00:11:01.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.787 "strip_size_kb": 0, 00:11:01.787 "state": "configuring", 00:11:01.787 "raid_level": "raid1", 00:11:01.787 "superblock": false, 00:11:01.787 "num_base_bdevs": 4, 00:11:01.787 "num_base_bdevs_discovered": 1, 00:11:01.787 "num_base_bdevs_operational": 4, 00:11:01.787 "base_bdevs_list": [ 00:11:01.787 { 00:11:01.787 "name": "BaseBdev1", 00:11:01.787 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:01.787 "is_configured": true, 00:11:01.787 "data_offset": 0, 00:11:01.787 "data_size": 65536 00:11:01.787 }, 00:11:01.787 { 00:11:01.787 "name": "BaseBdev2", 00:11:01.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.787 "is_configured": false, 00:11:01.787 "data_offset": 0, 00:11:01.787 "data_size": 0 00:11:01.787 }, 00:11:01.787 { 00:11:01.787 "name": "BaseBdev3", 00:11:01.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.787 "is_configured": false, 00:11:01.787 "data_offset": 0, 00:11:01.787 "data_size": 0 00:11:01.787 }, 00:11:01.787 { 00:11:01.787 "name": "BaseBdev4", 00:11:01.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.787 "is_configured": false, 00:11:01.787 "data_offset": 0, 00:11:01.787 "data_size": 0 00:11:01.787 } 00:11:01.787 ] 00:11:01.787 }' 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.787 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.046 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:02.046 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.046 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.046 [2024-10-29 11:00:07.542505] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:02.046 [2024-10-29 11:00:07.542578] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.305 [2024-10-29 11:00:07.554504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:02.305 [2024-10-29 11:00:07.556749] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:02.305 [2024-10-29 11:00:07.556793] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:02.305 [2024-10-29 11:00:07.556803] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:02.305 [2024-10-29 11:00:07.556811] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:02.305 [2024-10-29 11:00:07.556817] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:02.305 [2024-10-29 11:00:07.556826] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.305 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.305 "name": "Existed_Raid", 00:11:02.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.305 "strip_size_kb": 0, 00:11:02.305 "state": "configuring", 00:11:02.305 "raid_level": "raid1", 00:11:02.305 "superblock": false, 00:11:02.305 "num_base_bdevs": 4, 00:11:02.305 "num_base_bdevs_discovered": 1, 00:11:02.305 "num_base_bdevs_operational": 4, 00:11:02.305 "base_bdevs_list": [ 00:11:02.305 { 00:11:02.305 "name": "BaseBdev1", 00:11:02.305 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:02.305 "is_configured": true, 00:11:02.305 "data_offset": 0, 00:11:02.305 "data_size": 65536 00:11:02.305 }, 00:11:02.305 { 00:11:02.305 "name": "BaseBdev2", 00:11:02.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.305 "is_configured": false, 00:11:02.305 "data_offset": 0, 00:11:02.305 "data_size": 0 00:11:02.305 }, 00:11:02.305 { 00:11:02.305 "name": "BaseBdev3", 00:11:02.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.305 "is_configured": false, 00:11:02.305 "data_offset": 0, 00:11:02.305 "data_size": 0 00:11:02.305 }, 00:11:02.305 { 00:11:02.305 "name": "BaseBdev4", 00:11:02.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.305 "is_configured": false, 00:11:02.305 "data_offset": 0, 00:11:02.305 "data_size": 0 00:11:02.305 } 00:11:02.305 ] 00:11:02.305 }' 00:11:02.306 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.306 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.565 11:00:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:02.565 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.565 11:00:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.565 [2024-10-29 11:00:08.010392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:02.565 BaseBdev2 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.565 [ 00:11:02.565 { 00:11:02.565 "name": "BaseBdev2", 00:11:02.565 "aliases": [ 00:11:02.565 "792b298f-d5a2-400d-9298-2b6510d24308" 00:11:02.565 ], 00:11:02.565 "product_name": "Malloc disk", 00:11:02.565 "block_size": 512, 00:11:02.565 "num_blocks": 65536, 00:11:02.565 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:02.565 "assigned_rate_limits": { 00:11:02.565 "rw_ios_per_sec": 0, 00:11:02.565 "rw_mbytes_per_sec": 0, 00:11:02.565 "r_mbytes_per_sec": 0, 00:11:02.565 "w_mbytes_per_sec": 0 00:11:02.565 }, 00:11:02.565 "claimed": true, 00:11:02.565 "claim_type": "exclusive_write", 00:11:02.565 "zoned": false, 00:11:02.565 "supported_io_types": { 00:11:02.565 "read": true, 00:11:02.565 "write": true, 00:11:02.565 "unmap": true, 00:11:02.565 "flush": true, 00:11:02.565 "reset": true, 00:11:02.565 "nvme_admin": false, 00:11:02.565 "nvme_io": false, 00:11:02.565 "nvme_io_md": false, 00:11:02.565 "write_zeroes": true, 00:11:02.565 "zcopy": true, 00:11:02.565 "get_zone_info": false, 00:11:02.565 "zone_management": false, 00:11:02.565 "zone_append": false, 00:11:02.565 "compare": false, 00:11:02.565 "compare_and_write": false, 00:11:02.565 "abort": true, 00:11:02.565 "seek_hole": false, 00:11:02.565 "seek_data": false, 00:11:02.565 "copy": true, 00:11:02.565 "nvme_iov_md": false 00:11:02.565 }, 00:11:02.565 "memory_domains": [ 00:11:02.565 { 00:11:02.565 "dma_device_id": "system", 00:11:02.565 "dma_device_type": 1 00:11:02.565 }, 00:11:02.565 { 00:11:02.565 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.565 "dma_device_type": 2 00:11:02.565 } 00:11:02.565 ], 00:11:02.565 "driver_specific": {} 00:11:02.565 } 00:11:02.565 ] 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.565 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.825 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.825 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.825 "name": "Existed_Raid", 00:11:02.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.825 "strip_size_kb": 0, 00:11:02.825 "state": "configuring", 00:11:02.825 "raid_level": "raid1", 00:11:02.825 "superblock": false, 00:11:02.825 "num_base_bdevs": 4, 00:11:02.825 "num_base_bdevs_discovered": 2, 00:11:02.825 "num_base_bdevs_operational": 4, 00:11:02.825 "base_bdevs_list": [ 00:11:02.825 { 00:11:02.825 "name": "BaseBdev1", 00:11:02.825 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:02.825 "is_configured": true, 00:11:02.825 "data_offset": 0, 00:11:02.825 "data_size": 65536 00:11:02.825 }, 00:11:02.825 { 00:11:02.825 "name": "BaseBdev2", 00:11:02.825 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:02.825 "is_configured": true, 00:11:02.825 "data_offset": 0, 00:11:02.825 "data_size": 65536 00:11:02.825 }, 00:11:02.825 { 00:11:02.825 "name": "BaseBdev3", 00:11:02.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.825 "is_configured": false, 00:11:02.825 "data_offset": 0, 00:11:02.825 "data_size": 0 00:11:02.825 }, 00:11:02.825 { 00:11:02.825 "name": "BaseBdev4", 00:11:02.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.825 "is_configured": false, 00:11:02.825 "data_offset": 0, 00:11:02.825 "data_size": 0 00:11:02.825 } 00:11:02.825 ] 00:11:02.825 }' 00:11:02.825 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.825 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.085 [2024-10-29 11:00:08.513681] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:03.085 BaseBdev3 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.085 [ 00:11:03.085 { 00:11:03.085 "name": "BaseBdev3", 00:11:03.085 "aliases": [ 00:11:03.085 "b18fbfb1-23f8-457a-a644-8d2a8e64892b" 00:11:03.085 ], 00:11:03.085 "product_name": "Malloc disk", 00:11:03.085 "block_size": 512, 00:11:03.085 "num_blocks": 65536, 00:11:03.085 "uuid": "b18fbfb1-23f8-457a-a644-8d2a8e64892b", 00:11:03.085 "assigned_rate_limits": { 00:11:03.085 "rw_ios_per_sec": 0, 00:11:03.085 "rw_mbytes_per_sec": 0, 00:11:03.085 "r_mbytes_per_sec": 0, 00:11:03.085 "w_mbytes_per_sec": 0 00:11:03.085 }, 00:11:03.085 "claimed": true, 00:11:03.085 "claim_type": "exclusive_write", 00:11:03.085 "zoned": false, 00:11:03.085 "supported_io_types": { 00:11:03.085 "read": true, 00:11:03.085 "write": true, 00:11:03.085 "unmap": true, 00:11:03.085 "flush": true, 00:11:03.085 "reset": true, 00:11:03.085 "nvme_admin": false, 00:11:03.085 "nvme_io": false, 00:11:03.085 "nvme_io_md": false, 00:11:03.085 "write_zeroes": true, 00:11:03.085 "zcopy": true, 00:11:03.085 "get_zone_info": false, 00:11:03.085 "zone_management": false, 00:11:03.085 "zone_append": false, 00:11:03.085 "compare": false, 00:11:03.085 "compare_and_write": false, 00:11:03.085 "abort": true, 00:11:03.085 "seek_hole": false, 00:11:03.085 "seek_data": false, 00:11:03.085 "copy": true, 00:11:03.085 "nvme_iov_md": false 00:11:03.085 }, 00:11:03.085 "memory_domains": [ 00:11:03.085 { 00:11:03.085 "dma_device_id": "system", 00:11:03.085 "dma_device_type": 1 00:11:03.085 }, 00:11:03.085 { 00:11:03.085 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.085 "dma_device_type": 2 00:11:03.085 } 00:11:03.085 ], 00:11:03.085 "driver_specific": {} 00:11:03.085 } 00:11:03.085 ] 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.085 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.345 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.345 "name": "Existed_Raid", 00:11:03.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.345 "strip_size_kb": 0, 00:11:03.345 "state": "configuring", 00:11:03.345 "raid_level": "raid1", 00:11:03.345 "superblock": false, 00:11:03.345 "num_base_bdevs": 4, 00:11:03.345 "num_base_bdevs_discovered": 3, 00:11:03.345 "num_base_bdevs_operational": 4, 00:11:03.345 "base_bdevs_list": [ 00:11:03.345 { 00:11:03.345 "name": "BaseBdev1", 00:11:03.345 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:03.345 "is_configured": true, 00:11:03.345 "data_offset": 0, 00:11:03.345 "data_size": 65536 00:11:03.345 }, 00:11:03.345 { 00:11:03.345 "name": "BaseBdev2", 00:11:03.345 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:03.345 "is_configured": true, 00:11:03.345 "data_offset": 0, 00:11:03.345 "data_size": 65536 00:11:03.345 }, 00:11:03.345 { 00:11:03.345 "name": "BaseBdev3", 00:11:03.345 "uuid": "b18fbfb1-23f8-457a-a644-8d2a8e64892b", 00:11:03.345 "is_configured": true, 00:11:03.345 "data_offset": 0, 00:11:03.345 "data_size": 65536 00:11:03.345 }, 00:11:03.345 { 00:11:03.345 "name": "BaseBdev4", 00:11:03.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.345 "is_configured": false, 00:11:03.345 "data_offset": 0, 00:11:03.345 "data_size": 0 00:11:03.345 } 00:11:03.345 ] 00:11:03.345 }' 00:11:03.345 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.345 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.605 [2024-10-29 11:00:08.937932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:03.605 [2024-10-29 11:00:08.938006] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:11:03.605 [2024-10-29 11:00:08.938017] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:03.605 [2024-10-29 11:00:08.938351] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:11:03.605 [2024-10-29 11:00:08.938564] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:11:03.605 [2024-10-29 11:00:08.938585] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:11:03.605 [2024-10-29 11:00:08.938796] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.605 BaseBdev4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.605 [ 00:11:03.605 { 00:11:03.605 "name": "BaseBdev4", 00:11:03.605 "aliases": [ 00:11:03.605 "524e9065-e6c9-4282-b782-b5384fbbea70" 00:11:03.605 ], 00:11:03.605 "product_name": "Malloc disk", 00:11:03.605 "block_size": 512, 00:11:03.605 "num_blocks": 65536, 00:11:03.605 "uuid": "524e9065-e6c9-4282-b782-b5384fbbea70", 00:11:03.605 "assigned_rate_limits": { 00:11:03.605 "rw_ios_per_sec": 0, 00:11:03.605 "rw_mbytes_per_sec": 0, 00:11:03.605 "r_mbytes_per_sec": 0, 00:11:03.605 "w_mbytes_per_sec": 0 00:11:03.605 }, 00:11:03.605 "claimed": true, 00:11:03.605 "claim_type": "exclusive_write", 00:11:03.605 "zoned": false, 00:11:03.605 "supported_io_types": { 00:11:03.605 "read": true, 00:11:03.605 "write": true, 00:11:03.605 "unmap": true, 00:11:03.605 "flush": true, 00:11:03.605 "reset": true, 00:11:03.605 "nvme_admin": false, 00:11:03.605 "nvme_io": false, 00:11:03.605 "nvme_io_md": false, 00:11:03.605 "write_zeroes": true, 00:11:03.605 "zcopy": true, 00:11:03.605 "get_zone_info": false, 00:11:03.605 "zone_management": false, 00:11:03.605 "zone_append": false, 00:11:03.605 "compare": false, 00:11:03.605 "compare_and_write": false, 00:11:03.605 "abort": true, 00:11:03.605 "seek_hole": false, 00:11:03.605 "seek_data": false, 00:11:03.605 "copy": true, 00:11:03.605 "nvme_iov_md": false 00:11:03.605 }, 00:11:03.605 "memory_domains": [ 00:11:03.605 { 00:11:03.605 "dma_device_id": "system", 00:11:03.605 "dma_device_type": 1 00:11:03.605 }, 00:11:03.605 { 00:11:03.605 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.605 "dma_device_type": 2 00:11:03.605 } 00:11:03.605 ], 00:11:03.605 "driver_specific": {} 00:11:03.605 } 00:11:03.605 ] 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.605 11:00:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.605 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.605 "name": "Existed_Raid", 00:11:03.605 "uuid": "d259aa77-7e82-4a55-bd4c-78a9fcee3f36", 00:11:03.605 "strip_size_kb": 0, 00:11:03.605 "state": "online", 00:11:03.605 "raid_level": "raid1", 00:11:03.605 "superblock": false, 00:11:03.605 "num_base_bdevs": 4, 00:11:03.605 "num_base_bdevs_discovered": 4, 00:11:03.605 "num_base_bdevs_operational": 4, 00:11:03.605 "base_bdevs_list": [ 00:11:03.605 { 00:11:03.605 "name": "BaseBdev1", 00:11:03.605 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:03.605 "is_configured": true, 00:11:03.605 "data_offset": 0, 00:11:03.605 "data_size": 65536 00:11:03.605 }, 00:11:03.605 { 00:11:03.605 "name": "BaseBdev2", 00:11:03.605 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:03.605 "is_configured": true, 00:11:03.605 "data_offset": 0, 00:11:03.605 "data_size": 65536 00:11:03.605 }, 00:11:03.605 { 00:11:03.605 "name": "BaseBdev3", 00:11:03.605 "uuid": "b18fbfb1-23f8-457a-a644-8d2a8e64892b", 00:11:03.605 "is_configured": true, 00:11:03.605 "data_offset": 0, 00:11:03.605 "data_size": 65536 00:11:03.605 }, 00:11:03.605 { 00:11:03.605 "name": "BaseBdev4", 00:11:03.606 "uuid": "524e9065-e6c9-4282-b782-b5384fbbea70", 00:11:03.606 "is_configured": true, 00:11:03.606 "data_offset": 0, 00:11:03.606 "data_size": 65536 00:11:03.606 } 00:11:03.606 ] 00:11:03.606 }' 00:11:03.606 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.606 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:04.176 [2024-10-29 11:00:09.441494] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.176 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:04.176 "name": "Existed_Raid", 00:11:04.176 "aliases": [ 00:11:04.176 "d259aa77-7e82-4a55-bd4c-78a9fcee3f36" 00:11:04.176 ], 00:11:04.176 "product_name": "Raid Volume", 00:11:04.176 "block_size": 512, 00:11:04.176 "num_blocks": 65536, 00:11:04.176 "uuid": "d259aa77-7e82-4a55-bd4c-78a9fcee3f36", 00:11:04.176 "assigned_rate_limits": { 00:11:04.176 "rw_ios_per_sec": 0, 00:11:04.176 "rw_mbytes_per_sec": 0, 00:11:04.176 "r_mbytes_per_sec": 0, 00:11:04.176 "w_mbytes_per_sec": 0 00:11:04.176 }, 00:11:04.176 "claimed": false, 00:11:04.176 "zoned": false, 00:11:04.176 "supported_io_types": { 00:11:04.176 "read": true, 00:11:04.176 "write": true, 00:11:04.176 "unmap": false, 00:11:04.176 "flush": false, 00:11:04.176 "reset": true, 00:11:04.176 "nvme_admin": false, 00:11:04.176 "nvme_io": false, 00:11:04.176 "nvme_io_md": false, 00:11:04.176 "write_zeroes": true, 00:11:04.176 "zcopy": false, 00:11:04.176 "get_zone_info": false, 00:11:04.176 "zone_management": false, 00:11:04.176 "zone_append": false, 00:11:04.176 "compare": false, 00:11:04.177 "compare_and_write": false, 00:11:04.177 "abort": false, 00:11:04.177 "seek_hole": false, 00:11:04.177 "seek_data": false, 00:11:04.177 "copy": false, 00:11:04.177 "nvme_iov_md": false 00:11:04.177 }, 00:11:04.177 "memory_domains": [ 00:11:04.177 { 00:11:04.177 "dma_device_id": "system", 00:11:04.177 "dma_device_type": 1 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.177 "dma_device_type": 2 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "system", 00:11:04.177 "dma_device_type": 1 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.177 "dma_device_type": 2 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "system", 00:11:04.177 "dma_device_type": 1 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.177 "dma_device_type": 2 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "system", 00:11:04.177 "dma_device_type": 1 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.177 "dma_device_type": 2 00:11:04.177 } 00:11:04.177 ], 00:11:04.177 "driver_specific": { 00:11:04.177 "raid": { 00:11:04.177 "uuid": "d259aa77-7e82-4a55-bd4c-78a9fcee3f36", 00:11:04.177 "strip_size_kb": 0, 00:11:04.177 "state": "online", 00:11:04.177 "raid_level": "raid1", 00:11:04.177 "superblock": false, 00:11:04.177 "num_base_bdevs": 4, 00:11:04.177 "num_base_bdevs_discovered": 4, 00:11:04.177 "num_base_bdevs_operational": 4, 00:11:04.177 "base_bdevs_list": [ 00:11:04.177 { 00:11:04.177 "name": "BaseBdev1", 00:11:04.177 "uuid": "83f8c037-5b86-4cc6-8155-4ee7e9ad5590", 00:11:04.177 "is_configured": true, 00:11:04.177 "data_offset": 0, 00:11:04.177 "data_size": 65536 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "name": "BaseBdev2", 00:11:04.177 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:04.177 "is_configured": true, 00:11:04.177 "data_offset": 0, 00:11:04.177 "data_size": 65536 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "name": "BaseBdev3", 00:11:04.177 "uuid": "b18fbfb1-23f8-457a-a644-8d2a8e64892b", 00:11:04.177 "is_configured": true, 00:11:04.177 "data_offset": 0, 00:11:04.177 "data_size": 65536 00:11:04.177 }, 00:11:04.177 { 00:11:04.177 "name": "BaseBdev4", 00:11:04.177 "uuid": "524e9065-e6c9-4282-b782-b5384fbbea70", 00:11:04.177 "is_configured": true, 00:11:04.177 "data_offset": 0, 00:11:04.177 "data_size": 65536 00:11:04.177 } 00:11:04.177 ] 00:11:04.177 } 00:11:04.177 } 00:11:04.177 }' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:04.177 BaseBdev2 00:11:04.177 BaseBdev3 00:11:04.177 BaseBdev4' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.177 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.436 [2024-10-29 11:00:09.732736] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.436 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:04.436 "name": "Existed_Raid", 00:11:04.436 "uuid": "d259aa77-7e82-4a55-bd4c-78a9fcee3f36", 00:11:04.436 "strip_size_kb": 0, 00:11:04.436 "state": "online", 00:11:04.436 "raid_level": "raid1", 00:11:04.436 "superblock": false, 00:11:04.436 "num_base_bdevs": 4, 00:11:04.436 "num_base_bdevs_discovered": 3, 00:11:04.436 "num_base_bdevs_operational": 3, 00:11:04.436 "base_bdevs_list": [ 00:11:04.436 { 00:11:04.436 "name": null, 00:11:04.437 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:04.437 "is_configured": false, 00:11:04.437 "data_offset": 0, 00:11:04.437 "data_size": 65536 00:11:04.437 }, 00:11:04.437 { 00:11:04.437 "name": "BaseBdev2", 00:11:04.437 "uuid": "792b298f-d5a2-400d-9298-2b6510d24308", 00:11:04.437 "is_configured": true, 00:11:04.437 "data_offset": 0, 00:11:04.437 "data_size": 65536 00:11:04.437 }, 00:11:04.437 { 00:11:04.437 "name": "BaseBdev3", 00:11:04.437 "uuid": "b18fbfb1-23f8-457a-a644-8d2a8e64892b", 00:11:04.437 "is_configured": true, 00:11:04.437 "data_offset": 0, 00:11:04.437 "data_size": 65536 00:11:04.437 }, 00:11:04.437 { 00:11:04.437 "name": "BaseBdev4", 00:11:04.437 "uuid": "524e9065-e6c9-4282-b782-b5384fbbea70", 00:11:04.437 "is_configured": true, 00:11:04.437 "data_offset": 0, 00:11:04.437 "data_size": 65536 00:11:04.437 } 00:11:04.437 ] 00:11:04.437 }' 00:11:04.437 11:00:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:04.437 11:00:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.695 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 [2024-10-29 11:00:10.232638] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 [2024-10-29 11:00:10.304334] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 [2024-10-29 11:00:10.363750] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:04.955 [2024-10-29 11:00:10.363856] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:04.955 [2024-10-29 11:00:10.383766] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:04.955 [2024-10-29 11:00:10.383816] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:04.955 [2024-10-29 11:00:10.383830] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.955 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.215 BaseBdev2 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.215 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.215 [ 00:11:05.215 { 00:11:05.215 "name": "BaseBdev2", 00:11:05.215 "aliases": [ 00:11:05.215 "d42a6474-cb70-4555-9a30-505da81f63f8" 00:11:05.215 ], 00:11:05.215 "product_name": "Malloc disk", 00:11:05.215 "block_size": 512, 00:11:05.215 "num_blocks": 65536, 00:11:05.215 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:05.216 "assigned_rate_limits": { 00:11:05.216 "rw_ios_per_sec": 0, 00:11:05.216 "rw_mbytes_per_sec": 0, 00:11:05.216 "r_mbytes_per_sec": 0, 00:11:05.216 "w_mbytes_per_sec": 0 00:11:05.216 }, 00:11:05.216 "claimed": false, 00:11:05.216 "zoned": false, 00:11:05.216 "supported_io_types": { 00:11:05.216 "read": true, 00:11:05.216 "write": true, 00:11:05.216 "unmap": true, 00:11:05.216 "flush": true, 00:11:05.216 "reset": true, 00:11:05.216 "nvme_admin": false, 00:11:05.216 "nvme_io": false, 00:11:05.216 "nvme_io_md": false, 00:11:05.216 "write_zeroes": true, 00:11:05.216 "zcopy": true, 00:11:05.216 "get_zone_info": false, 00:11:05.216 "zone_management": false, 00:11:05.216 "zone_append": false, 00:11:05.216 "compare": false, 00:11:05.216 "compare_and_write": false, 00:11:05.216 "abort": true, 00:11:05.216 "seek_hole": false, 00:11:05.216 "seek_data": false, 00:11:05.216 "copy": true, 00:11:05.216 "nvme_iov_md": false 00:11:05.216 }, 00:11:05.216 "memory_domains": [ 00:11:05.216 { 00:11:05.216 "dma_device_id": "system", 00:11:05.216 "dma_device_type": 1 00:11:05.216 }, 00:11:05.216 { 00:11:05.216 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:05.216 "dma_device_type": 2 00:11:05.216 } 00:11:05.216 ], 00:11:05.216 "driver_specific": {} 00:11:05.216 } 00:11:05.216 ] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 BaseBdev3 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 [ 00:11:05.216 { 00:11:05.216 "name": "BaseBdev3", 00:11:05.216 "aliases": [ 00:11:05.216 "1103ccf6-14df-4c8b-9bfb-6d54040e10c1" 00:11:05.216 ], 00:11:05.216 "product_name": "Malloc disk", 00:11:05.216 "block_size": 512, 00:11:05.216 "num_blocks": 65536, 00:11:05.216 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:05.216 "assigned_rate_limits": { 00:11:05.216 "rw_ios_per_sec": 0, 00:11:05.216 "rw_mbytes_per_sec": 0, 00:11:05.216 "r_mbytes_per_sec": 0, 00:11:05.216 "w_mbytes_per_sec": 0 00:11:05.216 }, 00:11:05.216 "claimed": false, 00:11:05.216 "zoned": false, 00:11:05.216 "supported_io_types": { 00:11:05.216 "read": true, 00:11:05.216 "write": true, 00:11:05.216 "unmap": true, 00:11:05.216 "flush": true, 00:11:05.216 "reset": true, 00:11:05.216 "nvme_admin": false, 00:11:05.216 "nvme_io": false, 00:11:05.216 "nvme_io_md": false, 00:11:05.216 "write_zeroes": true, 00:11:05.216 "zcopy": true, 00:11:05.216 "get_zone_info": false, 00:11:05.216 "zone_management": false, 00:11:05.216 "zone_append": false, 00:11:05.216 "compare": false, 00:11:05.216 "compare_and_write": false, 00:11:05.216 "abort": true, 00:11:05.216 "seek_hole": false, 00:11:05.216 "seek_data": false, 00:11:05.216 "copy": true, 00:11:05.216 "nvme_iov_md": false 00:11:05.216 }, 00:11:05.216 "memory_domains": [ 00:11:05.216 { 00:11:05.216 "dma_device_id": "system", 00:11:05.216 "dma_device_type": 1 00:11:05.216 }, 00:11:05.216 { 00:11:05.216 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:05.216 "dma_device_type": 2 00:11:05.216 } 00:11:05.216 ], 00:11:05.216 "driver_specific": {} 00:11:05.216 } 00:11:05.216 ] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 BaseBdev4 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 [ 00:11:05.216 { 00:11:05.216 "name": "BaseBdev4", 00:11:05.216 "aliases": [ 00:11:05.216 "8098cc71-3a17-413c-a552-a3f48f3037e0" 00:11:05.216 ], 00:11:05.216 "product_name": "Malloc disk", 00:11:05.216 "block_size": 512, 00:11:05.216 "num_blocks": 65536, 00:11:05.216 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:05.216 "assigned_rate_limits": { 00:11:05.216 "rw_ios_per_sec": 0, 00:11:05.216 "rw_mbytes_per_sec": 0, 00:11:05.216 "r_mbytes_per_sec": 0, 00:11:05.216 "w_mbytes_per_sec": 0 00:11:05.216 }, 00:11:05.216 "claimed": false, 00:11:05.216 "zoned": false, 00:11:05.216 "supported_io_types": { 00:11:05.216 "read": true, 00:11:05.216 "write": true, 00:11:05.216 "unmap": true, 00:11:05.216 "flush": true, 00:11:05.216 "reset": true, 00:11:05.216 "nvme_admin": false, 00:11:05.216 "nvme_io": false, 00:11:05.216 "nvme_io_md": false, 00:11:05.216 "write_zeroes": true, 00:11:05.216 "zcopy": true, 00:11:05.216 "get_zone_info": false, 00:11:05.216 "zone_management": false, 00:11:05.216 "zone_append": false, 00:11:05.216 "compare": false, 00:11:05.216 "compare_and_write": false, 00:11:05.216 "abort": true, 00:11:05.216 "seek_hole": false, 00:11:05.216 "seek_data": false, 00:11:05.216 "copy": true, 00:11:05.216 "nvme_iov_md": false 00:11:05.216 }, 00:11:05.216 "memory_domains": [ 00:11:05.216 { 00:11:05.216 "dma_device_id": "system", 00:11:05.216 "dma_device_type": 1 00:11:05.216 }, 00:11:05.216 { 00:11:05.216 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:05.216 "dma_device_type": 2 00:11:05.216 } 00:11:05.216 ], 00:11:05.216 "driver_specific": {} 00:11:05.216 } 00:11:05.216 ] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.216 [2024-10-29 11:00:10.615795] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:05.216 [2024-10-29 11:00:10.615852] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:05.216 [2024-10-29 11:00:10.615873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:05.216 [2024-10-29 11:00:10.618026] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:05.216 [2024-10-29 11:00:10.618075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.216 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.217 "name": "Existed_Raid", 00:11:05.217 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.217 "strip_size_kb": 0, 00:11:05.217 "state": "configuring", 00:11:05.217 "raid_level": "raid1", 00:11:05.217 "superblock": false, 00:11:05.217 "num_base_bdevs": 4, 00:11:05.217 "num_base_bdevs_discovered": 3, 00:11:05.217 "num_base_bdevs_operational": 4, 00:11:05.217 "base_bdevs_list": [ 00:11:05.217 { 00:11:05.217 "name": "BaseBdev1", 00:11:05.217 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.217 "is_configured": false, 00:11:05.217 "data_offset": 0, 00:11:05.217 "data_size": 0 00:11:05.217 }, 00:11:05.217 { 00:11:05.217 "name": "BaseBdev2", 00:11:05.217 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:05.217 "is_configured": true, 00:11:05.217 "data_offset": 0, 00:11:05.217 "data_size": 65536 00:11:05.217 }, 00:11:05.217 { 00:11:05.217 "name": "BaseBdev3", 00:11:05.217 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:05.217 "is_configured": true, 00:11:05.217 "data_offset": 0, 00:11:05.217 "data_size": 65536 00:11:05.217 }, 00:11:05.217 { 00:11:05.217 "name": "BaseBdev4", 00:11:05.217 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:05.217 "is_configured": true, 00:11:05.217 "data_offset": 0, 00:11:05.217 "data_size": 65536 00:11:05.217 } 00:11:05.217 ] 00:11:05.217 }' 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.217 11:00:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.786 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:05.786 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.786 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.786 [2024-10-29 11:00:11.027201] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:05.786 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.787 "name": "Existed_Raid", 00:11:05.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.787 "strip_size_kb": 0, 00:11:05.787 "state": "configuring", 00:11:05.787 "raid_level": "raid1", 00:11:05.787 "superblock": false, 00:11:05.787 "num_base_bdevs": 4, 00:11:05.787 "num_base_bdevs_discovered": 2, 00:11:05.787 "num_base_bdevs_operational": 4, 00:11:05.787 "base_bdevs_list": [ 00:11:05.787 { 00:11:05.787 "name": "BaseBdev1", 00:11:05.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.787 "is_configured": false, 00:11:05.787 "data_offset": 0, 00:11:05.787 "data_size": 0 00:11:05.787 }, 00:11:05.787 { 00:11:05.787 "name": null, 00:11:05.787 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:05.787 "is_configured": false, 00:11:05.787 "data_offset": 0, 00:11:05.787 "data_size": 65536 00:11:05.787 }, 00:11:05.787 { 00:11:05.787 "name": "BaseBdev3", 00:11:05.787 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:05.787 "is_configured": true, 00:11:05.787 "data_offset": 0, 00:11:05.787 "data_size": 65536 00:11:05.787 }, 00:11:05.787 { 00:11:05.787 "name": "BaseBdev4", 00:11:05.787 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:05.787 "is_configured": true, 00:11:05.787 "data_offset": 0, 00:11:05.787 "data_size": 65536 00:11:05.787 } 00:11:05.787 ] 00:11:05.787 }' 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.787 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 [2024-10-29 11:00:11.483261] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:06.047 BaseBdev1 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 [ 00:11:06.047 { 00:11:06.047 "name": "BaseBdev1", 00:11:06.047 "aliases": [ 00:11:06.047 "47b181ff-acac-4b5b-9285-76a241f97400" 00:11:06.047 ], 00:11:06.047 "product_name": "Malloc disk", 00:11:06.047 "block_size": 512, 00:11:06.047 "num_blocks": 65536, 00:11:06.047 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:06.047 "assigned_rate_limits": { 00:11:06.047 "rw_ios_per_sec": 0, 00:11:06.047 "rw_mbytes_per_sec": 0, 00:11:06.047 "r_mbytes_per_sec": 0, 00:11:06.047 "w_mbytes_per_sec": 0 00:11:06.047 }, 00:11:06.047 "claimed": true, 00:11:06.047 "claim_type": "exclusive_write", 00:11:06.047 "zoned": false, 00:11:06.047 "supported_io_types": { 00:11:06.047 "read": true, 00:11:06.047 "write": true, 00:11:06.047 "unmap": true, 00:11:06.047 "flush": true, 00:11:06.047 "reset": true, 00:11:06.047 "nvme_admin": false, 00:11:06.047 "nvme_io": false, 00:11:06.047 "nvme_io_md": false, 00:11:06.047 "write_zeroes": true, 00:11:06.047 "zcopy": true, 00:11:06.047 "get_zone_info": false, 00:11:06.047 "zone_management": false, 00:11:06.047 "zone_append": false, 00:11:06.047 "compare": false, 00:11:06.047 "compare_and_write": false, 00:11:06.047 "abort": true, 00:11:06.047 "seek_hole": false, 00:11:06.047 "seek_data": false, 00:11:06.047 "copy": true, 00:11:06.047 "nvme_iov_md": false 00:11:06.047 }, 00:11:06.047 "memory_domains": [ 00:11:06.047 { 00:11:06.047 "dma_device_id": "system", 00:11:06.047 "dma_device_type": 1 00:11:06.047 }, 00:11:06.047 { 00:11:06.047 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:06.047 "dma_device_type": 2 00:11:06.047 } 00:11:06.047 ], 00:11:06.047 "driver_specific": {} 00:11:06.047 } 00:11:06.047 ] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.047 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.306 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.306 "name": "Existed_Raid", 00:11:06.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.306 "strip_size_kb": 0, 00:11:06.306 "state": "configuring", 00:11:06.306 "raid_level": "raid1", 00:11:06.306 "superblock": false, 00:11:06.306 "num_base_bdevs": 4, 00:11:06.306 "num_base_bdevs_discovered": 3, 00:11:06.306 "num_base_bdevs_operational": 4, 00:11:06.306 "base_bdevs_list": [ 00:11:06.306 { 00:11:06.306 "name": "BaseBdev1", 00:11:06.306 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:06.306 "is_configured": true, 00:11:06.306 "data_offset": 0, 00:11:06.306 "data_size": 65536 00:11:06.306 }, 00:11:06.306 { 00:11:06.306 "name": null, 00:11:06.306 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:06.306 "is_configured": false, 00:11:06.306 "data_offset": 0, 00:11:06.306 "data_size": 65536 00:11:06.306 }, 00:11:06.306 { 00:11:06.306 "name": "BaseBdev3", 00:11:06.306 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:06.306 "is_configured": true, 00:11:06.306 "data_offset": 0, 00:11:06.306 "data_size": 65536 00:11:06.306 }, 00:11:06.306 { 00:11:06.306 "name": "BaseBdev4", 00:11:06.306 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:06.306 "is_configured": true, 00:11:06.306 "data_offset": 0, 00:11:06.306 "data_size": 65536 00:11:06.306 } 00:11:06.306 ] 00:11:06.306 }' 00:11:06.306 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.306 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.565 [2024-10-29 11:00:11.978532] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.565 11:00:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.565 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.565 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.565 "name": "Existed_Raid", 00:11:06.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.565 "strip_size_kb": 0, 00:11:06.565 "state": "configuring", 00:11:06.565 "raid_level": "raid1", 00:11:06.565 "superblock": false, 00:11:06.565 "num_base_bdevs": 4, 00:11:06.565 "num_base_bdevs_discovered": 2, 00:11:06.565 "num_base_bdevs_operational": 4, 00:11:06.565 "base_bdevs_list": [ 00:11:06.565 { 00:11:06.565 "name": "BaseBdev1", 00:11:06.565 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:06.565 "is_configured": true, 00:11:06.565 "data_offset": 0, 00:11:06.565 "data_size": 65536 00:11:06.565 }, 00:11:06.565 { 00:11:06.565 "name": null, 00:11:06.565 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:06.565 "is_configured": false, 00:11:06.565 "data_offset": 0, 00:11:06.565 "data_size": 65536 00:11:06.565 }, 00:11:06.565 { 00:11:06.565 "name": null, 00:11:06.565 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:06.565 "is_configured": false, 00:11:06.565 "data_offset": 0, 00:11:06.565 "data_size": 65536 00:11:06.565 }, 00:11:06.565 { 00:11:06.565 "name": "BaseBdev4", 00:11:06.565 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:06.565 "is_configured": true, 00:11:06.565 "data_offset": 0, 00:11:06.565 "data_size": 65536 00:11:06.565 } 00:11:06.565 ] 00:11:06.565 }' 00:11:06.565 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.565 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.135 [2024-10-29 11:00:12.437881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.135 "name": "Existed_Raid", 00:11:07.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:07.135 "strip_size_kb": 0, 00:11:07.135 "state": "configuring", 00:11:07.135 "raid_level": "raid1", 00:11:07.135 "superblock": false, 00:11:07.135 "num_base_bdevs": 4, 00:11:07.135 "num_base_bdevs_discovered": 3, 00:11:07.135 "num_base_bdevs_operational": 4, 00:11:07.135 "base_bdevs_list": [ 00:11:07.135 { 00:11:07.135 "name": "BaseBdev1", 00:11:07.135 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:07.135 "is_configured": true, 00:11:07.135 "data_offset": 0, 00:11:07.135 "data_size": 65536 00:11:07.135 }, 00:11:07.135 { 00:11:07.135 "name": null, 00:11:07.135 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:07.135 "is_configured": false, 00:11:07.135 "data_offset": 0, 00:11:07.135 "data_size": 65536 00:11:07.135 }, 00:11:07.135 { 00:11:07.135 "name": "BaseBdev3", 00:11:07.135 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:07.135 "is_configured": true, 00:11:07.135 "data_offset": 0, 00:11:07.135 "data_size": 65536 00:11:07.135 }, 00:11:07.135 { 00:11:07.135 "name": "BaseBdev4", 00:11:07.135 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:07.135 "is_configured": true, 00:11:07.135 "data_offset": 0, 00:11:07.135 "data_size": 65536 00:11:07.135 } 00:11:07.135 ] 00:11:07.135 }' 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.135 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.395 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:07.395 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.395 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.395 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.655 [2024-10-29 11:00:12.921060] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.655 "name": "Existed_Raid", 00:11:07.655 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:07.655 "strip_size_kb": 0, 00:11:07.655 "state": "configuring", 00:11:07.655 "raid_level": "raid1", 00:11:07.655 "superblock": false, 00:11:07.655 "num_base_bdevs": 4, 00:11:07.655 "num_base_bdevs_discovered": 2, 00:11:07.655 "num_base_bdevs_operational": 4, 00:11:07.655 "base_bdevs_list": [ 00:11:07.655 { 00:11:07.655 "name": null, 00:11:07.655 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:07.655 "is_configured": false, 00:11:07.655 "data_offset": 0, 00:11:07.655 "data_size": 65536 00:11:07.655 }, 00:11:07.655 { 00:11:07.655 "name": null, 00:11:07.655 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:07.655 "is_configured": false, 00:11:07.655 "data_offset": 0, 00:11:07.655 "data_size": 65536 00:11:07.655 }, 00:11:07.655 { 00:11:07.655 "name": "BaseBdev3", 00:11:07.655 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:07.655 "is_configured": true, 00:11:07.655 "data_offset": 0, 00:11:07.655 "data_size": 65536 00:11:07.655 }, 00:11:07.655 { 00:11:07.655 "name": "BaseBdev4", 00:11:07.655 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:07.655 "is_configured": true, 00:11:07.655 "data_offset": 0, 00:11:07.655 "data_size": 65536 00:11:07.655 } 00:11:07.655 ] 00:11:07.655 }' 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.655 11:00:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:07.915 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.916 [2024-10-29 11:00:13.400346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.916 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.176 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.176 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.176 "name": "Existed_Raid", 00:11:08.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:08.176 "strip_size_kb": 0, 00:11:08.176 "state": "configuring", 00:11:08.176 "raid_level": "raid1", 00:11:08.176 "superblock": false, 00:11:08.176 "num_base_bdevs": 4, 00:11:08.176 "num_base_bdevs_discovered": 3, 00:11:08.176 "num_base_bdevs_operational": 4, 00:11:08.176 "base_bdevs_list": [ 00:11:08.176 { 00:11:08.176 "name": null, 00:11:08.176 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:08.176 "is_configured": false, 00:11:08.176 "data_offset": 0, 00:11:08.176 "data_size": 65536 00:11:08.176 }, 00:11:08.176 { 00:11:08.176 "name": "BaseBdev2", 00:11:08.176 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:08.176 "is_configured": true, 00:11:08.176 "data_offset": 0, 00:11:08.176 "data_size": 65536 00:11:08.176 }, 00:11:08.176 { 00:11:08.176 "name": "BaseBdev3", 00:11:08.176 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:08.176 "is_configured": true, 00:11:08.176 "data_offset": 0, 00:11:08.176 "data_size": 65536 00:11:08.176 }, 00:11:08.176 { 00:11:08.176 "name": "BaseBdev4", 00:11:08.176 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:08.176 "is_configured": true, 00:11:08.176 "data_offset": 0, 00:11:08.176 "data_size": 65536 00:11:08.176 } 00:11:08.176 ] 00:11:08.176 }' 00:11:08.176 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.176 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.435 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 47b181ff-acac-4b5b-9285-76a241f97400 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.436 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.695 [2024-10-29 11:00:13.952248] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:08.695 [2024-10-29 11:00:13.952423] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:11:08.695 [2024-10-29 11:00:13.952472] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:08.695 [2024-10-29 11:00:13.952844] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:11:08.695 [2024-10-29 11:00:13.953043] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:11:08.695 [2024-10-29 11:00:13.953085] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:11:08.695 [2024-10-29 11:00:13.953381] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.695 NewBaseBdev 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local i 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.695 [ 00:11:08.695 { 00:11:08.695 "name": "NewBaseBdev", 00:11:08.695 "aliases": [ 00:11:08.695 "47b181ff-acac-4b5b-9285-76a241f97400" 00:11:08.695 ], 00:11:08.695 "product_name": "Malloc disk", 00:11:08.695 "block_size": 512, 00:11:08.695 "num_blocks": 65536, 00:11:08.695 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:08.695 "assigned_rate_limits": { 00:11:08.695 "rw_ios_per_sec": 0, 00:11:08.695 "rw_mbytes_per_sec": 0, 00:11:08.695 "r_mbytes_per_sec": 0, 00:11:08.695 "w_mbytes_per_sec": 0 00:11:08.695 }, 00:11:08.695 "claimed": true, 00:11:08.695 "claim_type": "exclusive_write", 00:11:08.695 "zoned": false, 00:11:08.695 "supported_io_types": { 00:11:08.695 "read": true, 00:11:08.695 "write": true, 00:11:08.695 "unmap": true, 00:11:08.695 "flush": true, 00:11:08.695 "reset": true, 00:11:08.695 "nvme_admin": false, 00:11:08.695 "nvme_io": false, 00:11:08.695 "nvme_io_md": false, 00:11:08.695 "write_zeroes": true, 00:11:08.695 "zcopy": true, 00:11:08.695 "get_zone_info": false, 00:11:08.695 "zone_management": false, 00:11:08.695 "zone_append": false, 00:11:08.695 "compare": false, 00:11:08.695 "compare_and_write": false, 00:11:08.695 "abort": true, 00:11:08.695 "seek_hole": false, 00:11:08.695 "seek_data": false, 00:11:08.695 "copy": true, 00:11:08.695 "nvme_iov_md": false 00:11:08.695 }, 00:11:08.695 "memory_domains": [ 00:11:08.695 { 00:11:08.695 "dma_device_id": "system", 00:11:08.695 "dma_device_type": 1 00:11:08.695 }, 00:11:08.695 { 00:11:08.695 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.695 "dma_device_type": 2 00:11:08.695 } 00:11:08.695 ], 00:11:08.695 "driver_specific": {} 00:11:08.695 } 00:11:08.695 ] 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.695 11:00:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:08.695 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.695 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.695 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.695 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.695 "name": "Existed_Raid", 00:11:08.695 "uuid": "515ea77e-6c2b-4bf6-8a07-e7fedb1c86c4", 00:11:08.695 "strip_size_kb": 0, 00:11:08.695 "state": "online", 00:11:08.695 "raid_level": "raid1", 00:11:08.695 "superblock": false, 00:11:08.695 "num_base_bdevs": 4, 00:11:08.695 "num_base_bdevs_discovered": 4, 00:11:08.695 "num_base_bdevs_operational": 4, 00:11:08.695 "base_bdevs_list": [ 00:11:08.695 { 00:11:08.695 "name": "NewBaseBdev", 00:11:08.695 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:08.695 "is_configured": true, 00:11:08.695 "data_offset": 0, 00:11:08.695 "data_size": 65536 00:11:08.695 }, 00:11:08.695 { 00:11:08.695 "name": "BaseBdev2", 00:11:08.696 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:08.696 "is_configured": true, 00:11:08.696 "data_offset": 0, 00:11:08.696 "data_size": 65536 00:11:08.696 }, 00:11:08.696 { 00:11:08.696 "name": "BaseBdev3", 00:11:08.696 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:08.696 "is_configured": true, 00:11:08.696 "data_offset": 0, 00:11:08.696 "data_size": 65536 00:11:08.696 }, 00:11:08.696 { 00:11:08.696 "name": "BaseBdev4", 00:11:08.696 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:08.696 "is_configured": true, 00:11:08.696 "data_offset": 0, 00:11:08.696 "data_size": 65536 00:11:08.696 } 00:11:08.696 ] 00:11:08.696 }' 00:11:08.696 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.696 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.955 [2024-10-29 11:00:14.427946] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:08.955 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:09.215 "name": "Existed_Raid", 00:11:09.215 "aliases": [ 00:11:09.215 "515ea77e-6c2b-4bf6-8a07-e7fedb1c86c4" 00:11:09.215 ], 00:11:09.215 "product_name": "Raid Volume", 00:11:09.215 "block_size": 512, 00:11:09.215 "num_blocks": 65536, 00:11:09.215 "uuid": "515ea77e-6c2b-4bf6-8a07-e7fedb1c86c4", 00:11:09.215 "assigned_rate_limits": { 00:11:09.215 "rw_ios_per_sec": 0, 00:11:09.215 "rw_mbytes_per_sec": 0, 00:11:09.215 "r_mbytes_per_sec": 0, 00:11:09.215 "w_mbytes_per_sec": 0 00:11:09.215 }, 00:11:09.215 "claimed": false, 00:11:09.215 "zoned": false, 00:11:09.215 "supported_io_types": { 00:11:09.215 "read": true, 00:11:09.215 "write": true, 00:11:09.215 "unmap": false, 00:11:09.215 "flush": false, 00:11:09.215 "reset": true, 00:11:09.215 "nvme_admin": false, 00:11:09.215 "nvme_io": false, 00:11:09.215 "nvme_io_md": false, 00:11:09.215 "write_zeroes": true, 00:11:09.215 "zcopy": false, 00:11:09.215 "get_zone_info": false, 00:11:09.215 "zone_management": false, 00:11:09.215 "zone_append": false, 00:11:09.215 "compare": false, 00:11:09.215 "compare_and_write": false, 00:11:09.215 "abort": false, 00:11:09.215 "seek_hole": false, 00:11:09.215 "seek_data": false, 00:11:09.215 "copy": false, 00:11:09.215 "nvme_iov_md": false 00:11:09.215 }, 00:11:09.215 "memory_domains": [ 00:11:09.215 { 00:11:09.215 "dma_device_id": "system", 00:11:09.215 "dma_device_type": 1 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.215 "dma_device_type": 2 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "system", 00:11:09.215 "dma_device_type": 1 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.215 "dma_device_type": 2 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "system", 00:11:09.215 "dma_device_type": 1 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.215 "dma_device_type": 2 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "system", 00:11:09.215 "dma_device_type": 1 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:09.215 "dma_device_type": 2 00:11:09.215 } 00:11:09.215 ], 00:11:09.215 "driver_specific": { 00:11:09.215 "raid": { 00:11:09.215 "uuid": "515ea77e-6c2b-4bf6-8a07-e7fedb1c86c4", 00:11:09.215 "strip_size_kb": 0, 00:11:09.215 "state": "online", 00:11:09.215 "raid_level": "raid1", 00:11:09.215 "superblock": false, 00:11:09.215 "num_base_bdevs": 4, 00:11:09.215 "num_base_bdevs_discovered": 4, 00:11:09.215 "num_base_bdevs_operational": 4, 00:11:09.215 "base_bdevs_list": [ 00:11:09.215 { 00:11:09.215 "name": "NewBaseBdev", 00:11:09.215 "uuid": "47b181ff-acac-4b5b-9285-76a241f97400", 00:11:09.215 "is_configured": true, 00:11:09.215 "data_offset": 0, 00:11:09.215 "data_size": 65536 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "name": "BaseBdev2", 00:11:09.215 "uuid": "d42a6474-cb70-4555-9a30-505da81f63f8", 00:11:09.215 "is_configured": true, 00:11:09.215 "data_offset": 0, 00:11:09.215 "data_size": 65536 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "name": "BaseBdev3", 00:11:09.215 "uuid": "1103ccf6-14df-4c8b-9bfb-6d54040e10c1", 00:11:09.215 "is_configured": true, 00:11:09.215 "data_offset": 0, 00:11:09.215 "data_size": 65536 00:11:09.215 }, 00:11:09.215 { 00:11:09.215 "name": "BaseBdev4", 00:11:09.215 "uuid": "8098cc71-3a17-413c-a552-a3f48f3037e0", 00:11:09.215 "is_configured": true, 00:11:09.215 "data_offset": 0, 00:11:09.215 "data_size": 65536 00:11:09.215 } 00:11:09.215 ] 00:11:09.215 } 00:11:09.215 } 00:11:09.215 }' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:09.215 BaseBdev2 00:11:09.215 BaseBdev3 00:11:09.215 BaseBdev4' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.215 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.216 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.475 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:09.475 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:09.475 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:09.475 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:09.475 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.475 [2024-10-29 11:00:14.726994] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:09.475 [2024-10-29 11:00:14.727095] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:09.475 [2024-10-29 11:00:14.727239] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:09.476 [2024-10-29 11:00:14.727620] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:09.476 [2024-10-29 11:00:14.727693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 84135 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 84135 ']' 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # kill -0 84135 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # uname 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 84135 00:11:09.476 killing process with pid 84135 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 84135' 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@971 -- # kill 84135 00:11:09.476 [2024-10-29 11:00:14.775393] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:09.476 11:00:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@976 -- # wait 84135 00:11:09.476 [2024-10-29 11:00:14.856120] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:09.743 11:00:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:09.743 00:11:09.743 real 0m9.498s 00:11:09.743 user 0m15.956s 00:11:09.743 sys 0m2.066s 00:11:09.743 ************************************ 00:11:09.743 END TEST raid_state_function_test 00:11:09.743 ************************************ 00:11:09.743 11:00:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:09.743 11:00:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:10.019 11:00:15 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:11:10.019 11:00:15 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:11:10.019 11:00:15 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:10.019 11:00:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:10.019 ************************************ 00:11:10.019 START TEST raid_state_function_test_sb 00:11:10.019 ************************************ 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 4 true 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84790 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84790' 00:11:10.019 Process raid pid: 84790 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84790 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 84790 ']' 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:10.019 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:10.019 11:00:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.019 [2024-10-29 11:00:15.355040] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:10.019 [2024-10-29 11:00:15.355257] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:10.279 [2024-10-29 11:00:15.531777] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:10.279 [2024-10-29 11:00:15.574050] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:10.279 [2024-10-29 11:00:15.650166] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:10.279 [2024-10-29 11:00:15.650293] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.848 [2024-10-29 11:00:16.174574] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:10.848 [2024-10-29 11:00:16.174641] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:10.848 [2024-10-29 11:00:16.174664] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:10.848 [2024-10-29 11:00:16.174675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:10.848 [2024-10-29 11:00:16.174681] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:10.848 [2024-10-29 11:00:16.174693] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:10.848 [2024-10-29 11:00:16.174701] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:10.848 [2024-10-29 11:00:16.174710] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:10.848 "name": "Existed_Raid", 00:11:10.848 "uuid": "e0642c7e-ed72-4417-acf3-6b8e32cb098e", 00:11:10.848 "strip_size_kb": 0, 00:11:10.848 "state": "configuring", 00:11:10.848 "raid_level": "raid1", 00:11:10.848 "superblock": true, 00:11:10.848 "num_base_bdevs": 4, 00:11:10.848 "num_base_bdevs_discovered": 0, 00:11:10.848 "num_base_bdevs_operational": 4, 00:11:10.848 "base_bdevs_list": [ 00:11:10.848 { 00:11:10.848 "name": "BaseBdev1", 00:11:10.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.848 "is_configured": false, 00:11:10.848 "data_offset": 0, 00:11:10.848 "data_size": 0 00:11:10.848 }, 00:11:10.848 { 00:11:10.848 "name": "BaseBdev2", 00:11:10.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.848 "is_configured": false, 00:11:10.848 "data_offset": 0, 00:11:10.848 "data_size": 0 00:11:10.848 }, 00:11:10.848 { 00:11:10.848 "name": "BaseBdev3", 00:11:10.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.848 "is_configured": false, 00:11:10.848 "data_offset": 0, 00:11:10.848 "data_size": 0 00:11:10.848 }, 00:11:10.848 { 00:11:10.848 "name": "BaseBdev4", 00:11:10.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.848 "is_configured": false, 00:11:10.848 "data_offset": 0, 00:11:10.848 "data_size": 0 00:11:10.848 } 00:11:10.848 ] 00:11:10.848 }' 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:10.848 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 [2024-10-29 11:00:16.645656] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:11.419 [2024-10-29 11:00:16.645808] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 [2024-10-29 11:00:16.653612] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:11.419 [2024-10-29 11:00:16.653695] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:11.419 [2024-10-29 11:00:16.653722] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:11.419 [2024-10-29 11:00:16.653743] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:11.419 [2024-10-29 11:00:16.653759] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:11.419 [2024-10-29 11:00:16.653778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:11.419 [2024-10-29 11:00:16.653794] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:11.419 [2024-10-29 11:00:16.653814] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 [2024-10-29 11:00:16.676628] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:11.419 BaseBdev1 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.419 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.419 [ 00:11:11.419 { 00:11:11.419 "name": "BaseBdev1", 00:11:11.419 "aliases": [ 00:11:11.419 "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb" 00:11:11.419 ], 00:11:11.419 "product_name": "Malloc disk", 00:11:11.419 "block_size": 512, 00:11:11.420 "num_blocks": 65536, 00:11:11.420 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:11.420 "assigned_rate_limits": { 00:11:11.420 "rw_ios_per_sec": 0, 00:11:11.420 "rw_mbytes_per_sec": 0, 00:11:11.420 "r_mbytes_per_sec": 0, 00:11:11.420 "w_mbytes_per_sec": 0 00:11:11.420 }, 00:11:11.420 "claimed": true, 00:11:11.420 "claim_type": "exclusive_write", 00:11:11.420 "zoned": false, 00:11:11.420 "supported_io_types": { 00:11:11.420 "read": true, 00:11:11.420 "write": true, 00:11:11.420 "unmap": true, 00:11:11.420 "flush": true, 00:11:11.420 "reset": true, 00:11:11.420 "nvme_admin": false, 00:11:11.420 "nvme_io": false, 00:11:11.420 "nvme_io_md": false, 00:11:11.420 "write_zeroes": true, 00:11:11.420 "zcopy": true, 00:11:11.420 "get_zone_info": false, 00:11:11.420 "zone_management": false, 00:11:11.420 "zone_append": false, 00:11:11.420 "compare": false, 00:11:11.420 "compare_and_write": false, 00:11:11.420 "abort": true, 00:11:11.420 "seek_hole": false, 00:11:11.420 "seek_data": false, 00:11:11.420 "copy": true, 00:11:11.420 "nvme_iov_md": false 00:11:11.420 }, 00:11:11.420 "memory_domains": [ 00:11:11.420 { 00:11:11.420 "dma_device_id": "system", 00:11:11.420 "dma_device_type": 1 00:11:11.420 }, 00:11:11.420 { 00:11:11.420 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:11.420 "dma_device_type": 2 00:11:11.420 } 00:11:11.420 ], 00:11:11.420 "driver_specific": {} 00:11:11.420 } 00:11:11.420 ] 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.420 "name": "Existed_Raid", 00:11:11.420 "uuid": "18297b12-a53c-41a4-9414-9ba84355978a", 00:11:11.420 "strip_size_kb": 0, 00:11:11.420 "state": "configuring", 00:11:11.420 "raid_level": "raid1", 00:11:11.420 "superblock": true, 00:11:11.420 "num_base_bdevs": 4, 00:11:11.420 "num_base_bdevs_discovered": 1, 00:11:11.420 "num_base_bdevs_operational": 4, 00:11:11.420 "base_bdevs_list": [ 00:11:11.420 { 00:11:11.420 "name": "BaseBdev1", 00:11:11.420 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:11.420 "is_configured": true, 00:11:11.420 "data_offset": 2048, 00:11:11.420 "data_size": 63488 00:11:11.420 }, 00:11:11.420 { 00:11:11.420 "name": "BaseBdev2", 00:11:11.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.420 "is_configured": false, 00:11:11.420 "data_offset": 0, 00:11:11.420 "data_size": 0 00:11:11.420 }, 00:11:11.420 { 00:11:11.420 "name": "BaseBdev3", 00:11:11.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.420 "is_configured": false, 00:11:11.420 "data_offset": 0, 00:11:11.420 "data_size": 0 00:11:11.420 }, 00:11:11.420 { 00:11:11.420 "name": "BaseBdev4", 00:11:11.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.420 "is_configured": false, 00:11:11.420 "data_offset": 0, 00:11:11.420 "data_size": 0 00:11:11.420 } 00:11:11.420 ] 00:11:11.420 }' 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.420 11:00:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.680 [2024-10-29 11:00:17.151925] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:11.680 [2024-10-29 11:00:17.152109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.680 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.680 [2024-10-29 11:00:17.163905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:11.680 [2024-10-29 11:00:17.166185] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:11.681 [2024-10-29 11:00:17.166236] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:11.681 [2024-10-29 11:00:17.166246] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:11.681 [2024-10-29 11:00:17.166255] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:11.681 [2024-10-29 11:00:17.166262] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:11.681 [2024-10-29 11:00:17.166270] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.681 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.941 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.941 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.941 "name": "Existed_Raid", 00:11:11.941 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:11.941 "strip_size_kb": 0, 00:11:11.941 "state": "configuring", 00:11:11.941 "raid_level": "raid1", 00:11:11.941 "superblock": true, 00:11:11.941 "num_base_bdevs": 4, 00:11:11.941 "num_base_bdevs_discovered": 1, 00:11:11.941 "num_base_bdevs_operational": 4, 00:11:11.941 "base_bdevs_list": [ 00:11:11.941 { 00:11:11.941 "name": "BaseBdev1", 00:11:11.941 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:11.941 "is_configured": true, 00:11:11.941 "data_offset": 2048, 00:11:11.941 "data_size": 63488 00:11:11.941 }, 00:11:11.941 { 00:11:11.941 "name": "BaseBdev2", 00:11:11.941 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.941 "is_configured": false, 00:11:11.941 "data_offset": 0, 00:11:11.941 "data_size": 0 00:11:11.941 }, 00:11:11.941 { 00:11:11.941 "name": "BaseBdev3", 00:11:11.941 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.941 "is_configured": false, 00:11:11.941 "data_offset": 0, 00:11:11.941 "data_size": 0 00:11:11.941 }, 00:11:11.941 { 00:11:11.941 "name": "BaseBdev4", 00:11:11.941 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.941 "is_configured": false, 00:11:11.941 "data_offset": 0, 00:11:11.941 "data_size": 0 00:11:11.941 } 00:11:11.941 ] 00:11:11.941 }' 00:11:11.941 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.941 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.202 [2024-10-29 11:00:17.544970] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:12.202 BaseBdev2 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.202 [ 00:11:12.202 { 00:11:12.202 "name": "BaseBdev2", 00:11:12.202 "aliases": [ 00:11:12.202 "3f475558-6e55-444f-87fd-551fba722f45" 00:11:12.202 ], 00:11:12.202 "product_name": "Malloc disk", 00:11:12.202 "block_size": 512, 00:11:12.202 "num_blocks": 65536, 00:11:12.202 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:12.202 "assigned_rate_limits": { 00:11:12.202 "rw_ios_per_sec": 0, 00:11:12.202 "rw_mbytes_per_sec": 0, 00:11:12.202 "r_mbytes_per_sec": 0, 00:11:12.202 "w_mbytes_per_sec": 0 00:11:12.202 }, 00:11:12.202 "claimed": true, 00:11:12.202 "claim_type": "exclusive_write", 00:11:12.202 "zoned": false, 00:11:12.202 "supported_io_types": { 00:11:12.202 "read": true, 00:11:12.202 "write": true, 00:11:12.202 "unmap": true, 00:11:12.202 "flush": true, 00:11:12.202 "reset": true, 00:11:12.202 "nvme_admin": false, 00:11:12.202 "nvme_io": false, 00:11:12.202 "nvme_io_md": false, 00:11:12.202 "write_zeroes": true, 00:11:12.202 "zcopy": true, 00:11:12.202 "get_zone_info": false, 00:11:12.202 "zone_management": false, 00:11:12.202 "zone_append": false, 00:11:12.202 "compare": false, 00:11:12.202 "compare_and_write": false, 00:11:12.202 "abort": true, 00:11:12.202 "seek_hole": false, 00:11:12.202 "seek_data": false, 00:11:12.202 "copy": true, 00:11:12.202 "nvme_iov_md": false 00:11:12.202 }, 00:11:12.202 "memory_domains": [ 00:11:12.202 { 00:11:12.202 "dma_device_id": "system", 00:11:12.202 "dma_device_type": 1 00:11:12.202 }, 00:11:12.202 { 00:11:12.202 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:12.202 "dma_device_type": 2 00:11:12.202 } 00:11:12.202 ], 00:11:12.202 "driver_specific": {} 00:11:12.202 } 00:11:12.202 ] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.202 "name": "Existed_Raid", 00:11:12.202 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:12.202 "strip_size_kb": 0, 00:11:12.202 "state": "configuring", 00:11:12.202 "raid_level": "raid1", 00:11:12.202 "superblock": true, 00:11:12.202 "num_base_bdevs": 4, 00:11:12.202 "num_base_bdevs_discovered": 2, 00:11:12.202 "num_base_bdevs_operational": 4, 00:11:12.202 "base_bdevs_list": [ 00:11:12.202 { 00:11:12.202 "name": "BaseBdev1", 00:11:12.202 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:12.202 "is_configured": true, 00:11:12.202 "data_offset": 2048, 00:11:12.202 "data_size": 63488 00:11:12.202 }, 00:11:12.202 { 00:11:12.202 "name": "BaseBdev2", 00:11:12.202 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:12.202 "is_configured": true, 00:11:12.202 "data_offset": 2048, 00:11:12.202 "data_size": 63488 00:11:12.202 }, 00:11:12.202 { 00:11:12.202 "name": "BaseBdev3", 00:11:12.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.202 "is_configured": false, 00:11:12.202 "data_offset": 0, 00:11:12.202 "data_size": 0 00:11:12.202 }, 00:11:12.202 { 00:11:12.202 "name": "BaseBdev4", 00:11:12.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.202 "is_configured": false, 00:11:12.202 "data_offset": 0, 00:11:12.202 "data_size": 0 00:11:12.202 } 00:11:12.202 ] 00:11:12.202 }' 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.202 11:00:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.773 [2024-10-29 11:00:18.056951] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:12.773 BaseBdev3 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.773 [ 00:11:12.773 { 00:11:12.773 "name": "BaseBdev3", 00:11:12.773 "aliases": [ 00:11:12.773 "b4b5c326-40eb-4089-a0a3-5fa1ca64558d" 00:11:12.773 ], 00:11:12.773 "product_name": "Malloc disk", 00:11:12.773 "block_size": 512, 00:11:12.773 "num_blocks": 65536, 00:11:12.773 "uuid": "b4b5c326-40eb-4089-a0a3-5fa1ca64558d", 00:11:12.773 "assigned_rate_limits": { 00:11:12.773 "rw_ios_per_sec": 0, 00:11:12.773 "rw_mbytes_per_sec": 0, 00:11:12.773 "r_mbytes_per_sec": 0, 00:11:12.773 "w_mbytes_per_sec": 0 00:11:12.773 }, 00:11:12.773 "claimed": true, 00:11:12.773 "claim_type": "exclusive_write", 00:11:12.773 "zoned": false, 00:11:12.773 "supported_io_types": { 00:11:12.773 "read": true, 00:11:12.773 "write": true, 00:11:12.773 "unmap": true, 00:11:12.773 "flush": true, 00:11:12.773 "reset": true, 00:11:12.773 "nvme_admin": false, 00:11:12.773 "nvme_io": false, 00:11:12.773 "nvme_io_md": false, 00:11:12.773 "write_zeroes": true, 00:11:12.773 "zcopy": true, 00:11:12.773 "get_zone_info": false, 00:11:12.773 "zone_management": false, 00:11:12.773 "zone_append": false, 00:11:12.773 "compare": false, 00:11:12.773 "compare_and_write": false, 00:11:12.773 "abort": true, 00:11:12.773 "seek_hole": false, 00:11:12.773 "seek_data": false, 00:11:12.773 "copy": true, 00:11:12.773 "nvme_iov_md": false 00:11:12.773 }, 00:11:12.773 "memory_domains": [ 00:11:12.773 { 00:11:12.773 "dma_device_id": "system", 00:11:12.773 "dma_device_type": 1 00:11:12.773 }, 00:11:12.773 { 00:11:12.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:12.773 "dma_device_type": 2 00:11:12.773 } 00:11:12.773 ], 00:11:12.773 "driver_specific": {} 00:11:12.773 } 00:11:12.773 ] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.773 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.773 "name": "Existed_Raid", 00:11:12.773 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:12.773 "strip_size_kb": 0, 00:11:12.773 "state": "configuring", 00:11:12.773 "raid_level": "raid1", 00:11:12.773 "superblock": true, 00:11:12.774 "num_base_bdevs": 4, 00:11:12.774 "num_base_bdevs_discovered": 3, 00:11:12.774 "num_base_bdevs_operational": 4, 00:11:12.774 "base_bdevs_list": [ 00:11:12.774 { 00:11:12.774 "name": "BaseBdev1", 00:11:12.774 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:12.774 "is_configured": true, 00:11:12.774 "data_offset": 2048, 00:11:12.774 "data_size": 63488 00:11:12.774 }, 00:11:12.774 { 00:11:12.774 "name": "BaseBdev2", 00:11:12.774 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:12.774 "is_configured": true, 00:11:12.774 "data_offset": 2048, 00:11:12.774 "data_size": 63488 00:11:12.774 }, 00:11:12.774 { 00:11:12.774 "name": "BaseBdev3", 00:11:12.774 "uuid": "b4b5c326-40eb-4089-a0a3-5fa1ca64558d", 00:11:12.774 "is_configured": true, 00:11:12.774 "data_offset": 2048, 00:11:12.774 "data_size": 63488 00:11:12.774 }, 00:11:12.774 { 00:11:12.774 "name": "BaseBdev4", 00:11:12.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.774 "is_configured": false, 00:11:12.774 "data_offset": 0, 00:11:12.774 "data_size": 0 00:11:12.774 } 00:11:12.774 ] 00:11:12.774 }' 00:11:12.774 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.774 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.034 [2024-10-29 11:00:18.485891] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:13.034 [2024-10-29 11:00:18.486317] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:11:13.034 [2024-10-29 11:00:18.486398] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:13.034 [2024-10-29 11:00:18.486793] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:11:13.034 BaseBdev4 00:11:13.034 [2024-10-29 11:00:18.487018] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:11:13.034 [2024-10-29 11:00:18.487083] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:11:13.034 [2024-10-29 11:00:18.487307] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.034 [ 00:11:13.034 { 00:11:13.034 "name": "BaseBdev4", 00:11:13.034 "aliases": [ 00:11:13.034 "be1b5c98-1a99-4922-9330-d1ba8c2e19e6" 00:11:13.034 ], 00:11:13.034 "product_name": "Malloc disk", 00:11:13.034 "block_size": 512, 00:11:13.034 "num_blocks": 65536, 00:11:13.034 "uuid": "be1b5c98-1a99-4922-9330-d1ba8c2e19e6", 00:11:13.034 "assigned_rate_limits": { 00:11:13.034 "rw_ios_per_sec": 0, 00:11:13.034 "rw_mbytes_per_sec": 0, 00:11:13.034 "r_mbytes_per_sec": 0, 00:11:13.034 "w_mbytes_per_sec": 0 00:11:13.034 }, 00:11:13.034 "claimed": true, 00:11:13.034 "claim_type": "exclusive_write", 00:11:13.034 "zoned": false, 00:11:13.034 "supported_io_types": { 00:11:13.034 "read": true, 00:11:13.034 "write": true, 00:11:13.034 "unmap": true, 00:11:13.034 "flush": true, 00:11:13.034 "reset": true, 00:11:13.034 "nvme_admin": false, 00:11:13.034 "nvme_io": false, 00:11:13.034 "nvme_io_md": false, 00:11:13.034 "write_zeroes": true, 00:11:13.034 "zcopy": true, 00:11:13.034 "get_zone_info": false, 00:11:13.034 "zone_management": false, 00:11:13.034 "zone_append": false, 00:11:13.034 "compare": false, 00:11:13.034 "compare_and_write": false, 00:11:13.034 "abort": true, 00:11:13.034 "seek_hole": false, 00:11:13.034 "seek_data": false, 00:11:13.034 "copy": true, 00:11:13.034 "nvme_iov_md": false 00:11:13.034 }, 00:11:13.034 "memory_domains": [ 00:11:13.034 { 00:11:13.034 "dma_device_id": "system", 00:11:13.034 "dma_device_type": 1 00:11:13.034 }, 00:11:13.034 { 00:11:13.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.034 "dma_device_type": 2 00:11:13.034 } 00:11:13.034 ], 00:11:13.034 "driver_specific": {} 00:11:13.034 } 00:11:13.034 ] 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.034 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.035 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.295 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.295 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.295 "name": "Existed_Raid", 00:11:13.295 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:13.295 "strip_size_kb": 0, 00:11:13.295 "state": "online", 00:11:13.295 "raid_level": "raid1", 00:11:13.295 "superblock": true, 00:11:13.295 "num_base_bdevs": 4, 00:11:13.295 "num_base_bdevs_discovered": 4, 00:11:13.295 "num_base_bdevs_operational": 4, 00:11:13.295 "base_bdevs_list": [ 00:11:13.295 { 00:11:13.295 "name": "BaseBdev1", 00:11:13.295 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:13.295 "is_configured": true, 00:11:13.295 "data_offset": 2048, 00:11:13.295 "data_size": 63488 00:11:13.295 }, 00:11:13.295 { 00:11:13.295 "name": "BaseBdev2", 00:11:13.295 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:13.295 "is_configured": true, 00:11:13.295 "data_offset": 2048, 00:11:13.295 "data_size": 63488 00:11:13.295 }, 00:11:13.295 { 00:11:13.295 "name": "BaseBdev3", 00:11:13.295 "uuid": "b4b5c326-40eb-4089-a0a3-5fa1ca64558d", 00:11:13.295 "is_configured": true, 00:11:13.295 "data_offset": 2048, 00:11:13.295 "data_size": 63488 00:11:13.295 }, 00:11:13.295 { 00:11:13.295 "name": "BaseBdev4", 00:11:13.295 "uuid": "be1b5c98-1a99-4922-9330-d1ba8c2e19e6", 00:11:13.295 "is_configured": true, 00:11:13.295 "data_offset": 2048, 00:11:13.295 "data_size": 63488 00:11:13.295 } 00:11:13.295 ] 00:11:13.295 }' 00:11:13.295 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.295 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.555 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.556 [2024-10-29 11:00:18.969589] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:13.556 11:00:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.556 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:13.556 "name": "Existed_Raid", 00:11:13.556 "aliases": [ 00:11:13.556 "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa" 00:11:13.556 ], 00:11:13.556 "product_name": "Raid Volume", 00:11:13.556 "block_size": 512, 00:11:13.556 "num_blocks": 63488, 00:11:13.556 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:13.556 "assigned_rate_limits": { 00:11:13.556 "rw_ios_per_sec": 0, 00:11:13.556 "rw_mbytes_per_sec": 0, 00:11:13.556 "r_mbytes_per_sec": 0, 00:11:13.556 "w_mbytes_per_sec": 0 00:11:13.556 }, 00:11:13.556 "claimed": false, 00:11:13.556 "zoned": false, 00:11:13.556 "supported_io_types": { 00:11:13.556 "read": true, 00:11:13.556 "write": true, 00:11:13.556 "unmap": false, 00:11:13.556 "flush": false, 00:11:13.556 "reset": true, 00:11:13.556 "nvme_admin": false, 00:11:13.556 "nvme_io": false, 00:11:13.556 "nvme_io_md": false, 00:11:13.556 "write_zeroes": true, 00:11:13.556 "zcopy": false, 00:11:13.556 "get_zone_info": false, 00:11:13.556 "zone_management": false, 00:11:13.556 "zone_append": false, 00:11:13.556 "compare": false, 00:11:13.556 "compare_and_write": false, 00:11:13.556 "abort": false, 00:11:13.556 "seek_hole": false, 00:11:13.556 "seek_data": false, 00:11:13.556 "copy": false, 00:11:13.556 "nvme_iov_md": false 00:11:13.556 }, 00:11:13.556 "memory_domains": [ 00:11:13.556 { 00:11:13.556 "dma_device_id": "system", 00:11:13.556 "dma_device_type": 1 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.556 "dma_device_type": 2 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "system", 00:11:13.556 "dma_device_type": 1 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.556 "dma_device_type": 2 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "system", 00:11:13.556 "dma_device_type": 1 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.556 "dma_device_type": 2 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "system", 00:11:13.556 "dma_device_type": 1 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.556 "dma_device_type": 2 00:11:13.556 } 00:11:13.556 ], 00:11:13.556 "driver_specific": { 00:11:13.556 "raid": { 00:11:13.556 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:13.556 "strip_size_kb": 0, 00:11:13.556 "state": "online", 00:11:13.556 "raid_level": "raid1", 00:11:13.556 "superblock": true, 00:11:13.556 "num_base_bdevs": 4, 00:11:13.556 "num_base_bdevs_discovered": 4, 00:11:13.556 "num_base_bdevs_operational": 4, 00:11:13.556 "base_bdevs_list": [ 00:11:13.556 { 00:11:13.556 "name": "BaseBdev1", 00:11:13.556 "uuid": "8dc4937c-2ab2-4bc6-b026-fd38b29bcffb", 00:11:13.556 "is_configured": true, 00:11:13.556 "data_offset": 2048, 00:11:13.556 "data_size": 63488 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "name": "BaseBdev2", 00:11:13.556 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:13.556 "is_configured": true, 00:11:13.556 "data_offset": 2048, 00:11:13.556 "data_size": 63488 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "name": "BaseBdev3", 00:11:13.556 "uuid": "b4b5c326-40eb-4089-a0a3-5fa1ca64558d", 00:11:13.556 "is_configured": true, 00:11:13.556 "data_offset": 2048, 00:11:13.556 "data_size": 63488 00:11:13.556 }, 00:11:13.556 { 00:11:13.556 "name": "BaseBdev4", 00:11:13.556 "uuid": "be1b5c98-1a99-4922-9330-d1ba8c2e19e6", 00:11:13.556 "is_configured": true, 00:11:13.556 "data_offset": 2048, 00:11:13.556 "data_size": 63488 00:11:13.556 } 00:11:13.556 ] 00:11:13.556 } 00:11:13.556 } 00:11:13.556 }' 00:11:13.556 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:13.816 BaseBdev2 00:11:13.816 BaseBdev3 00:11:13.816 BaseBdev4' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.816 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.817 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.817 [2024-10-29 11:00:19.296755] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.076 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.076 "name": "Existed_Raid", 00:11:14.076 "uuid": "b88b29c5-a32d-4a76-b7aa-d4c25e165cfa", 00:11:14.076 "strip_size_kb": 0, 00:11:14.076 "state": "online", 00:11:14.076 "raid_level": "raid1", 00:11:14.076 "superblock": true, 00:11:14.076 "num_base_bdevs": 4, 00:11:14.076 "num_base_bdevs_discovered": 3, 00:11:14.076 "num_base_bdevs_operational": 3, 00:11:14.076 "base_bdevs_list": [ 00:11:14.076 { 00:11:14.076 "name": null, 00:11:14.076 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.076 "is_configured": false, 00:11:14.076 "data_offset": 0, 00:11:14.076 "data_size": 63488 00:11:14.076 }, 00:11:14.077 { 00:11:14.077 "name": "BaseBdev2", 00:11:14.077 "uuid": "3f475558-6e55-444f-87fd-551fba722f45", 00:11:14.077 "is_configured": true, 00:11:14.077 "data_offset": 2048, 00:11:14.077 "data_size": 63488 00:11:14.077 }, 00:11:14.077 { 00:11:14.077 "name": "BaseBdev3", 00:11:14.077 "uuid": "b4b5c326-40eb-4089-a0a3-5fa1ca64558d", 00:11:14.077 "is_configured": true, 00:11:14.077 "data_offset": 2048, 00:11:14.077 "data_size": 63488 00:11:14.077 }, 00:11:14.077 { 00:11:14.077 "name": "BaseBdev4", 00:11:14.077 "uuid": "be1b5c98-1a99-4922-9330-d1ba8c2e19e6", 00:11:14.077 "is_configured": true, 00:11:14.077 "data_offset": 2048, 00:11:14.077 "data_size": 63488 00:11:14.077 } 00:11:14.077 ] 00:11:14.077 }' 00:11:14.077 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.077 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.335 [2024-10-29 11:00:19.777221] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.335 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 [2024-10-29 11:00:19.853857] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 [2024-10-29 11:00:19.934335] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:14.595 [2024-10-29 11:00:19.934571] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:14.595 [2024-10-29 11:00:19.955878] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:14.595 [2024-10-29 11:00:19.956021] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:14.595 [2024-10-29 11:00:19.956073] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 11:00:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 BaseBdev2 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.595 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.595 [ 00:11:14.595 { 00:11:14.595 "name": "BaseBdev2", 00:11:14.595 "aliases": [ 00:11:14.595 "afa6577d-5918-4b99-84bb-8c66400cfd33" 00:11:14.595 ], 00:11:14.595 "product_name": "Malloc disk", 00:11:14.595 "block_size": 512, 00:11:14.595 "num_blocks": 65536, 00:11:14.595 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:14.595 "assigned_rate_limits": { 00:11:14.595 "rw_ios_per_sec": 0, 00:11:14.595 "rw_mbytes_per_sec": 0, 00:11:14.595 "r_mbytes_per_sec": 0, 00:11:14.595 "w_mbytes_per_sec": 0 00:11:14.595 }, 00:11:14.596 "claimed": false, 00:11:14.596 "zoned": false, 00:11:14.596 "supported_io_types": { 00:11:14.596 "read": true, 00:11:14.596 "write": true, 00:11:14.596 "unmap": true, 00:11:14.596 "flush": true, 00:11:14.596 "reset": true, 00:11:14.596 "nvme_admin": false, 00:11:14.596 "nvme_io": false, 00:11:14.596 "nvme_io_md": false, 00:11:14.596 "write_zeroes": true, 00:11:14.596 "zcopy": true, 00:11:14.596 "get_zone_info": false, 00:11:14.596 "zone_management": false, 00:11:14.596 "zone_append": false, 00:11:14.596 "compare": false, 00:11:14.596 "compare_and_write": false, 00:11:14.596 "abort": true, 00:11:14.596 "seek_hole": false, 00:11:14.596 "seek_data": false, 00:11:14.596 "copy": true, 00:11:14.596 "nvme_iov_md": false 00:11:14.596 }, 00:11:14.596 "memory_domains": [ 00:11:14.596 { 00:11:14.596 "dma_device_id": "system", 00:11:14.596 "dma_device_type": 1 00:11:14.596 }, 00:11:14.596 { 00:11:14.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.596 "dma_device_type": 2 00:11:14.596 } 00:11:14.596 ], 00:11:14.596 "driver_specific": {} 00:11:14.596 } 00:11:14.596 ] 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.596 BaseBdev3 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.596 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.856 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.856 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:14.856 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.856 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.856 [ 00:11:14.856 { 00:11:14.856 "name": "BaseBdev3", 00:11:14.856 "aliases": [ 00:11:14.857 "f4e84ee4-88c4-4cfa-9933-b41071172319" 00:11:14.857 ], 00:11:14.857 "product_name": "Malloc disk", 00:11:14.857 "block_size": 512, 00:11:14.857 "num_blocks": 65536, 00:11:14.857 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:14.857 "assigned_rate_limits": { 00:11:14.857 "rw_ios_per_sec": 0, 00:11:14.857 "rw_mbytes_per_sec": 0, 00:11:14.857 "r_mbytes_per_sec": 0, 00:11:14.857 "w_mbytes_per_sec": 0 00:11:14.857 }, 00:11:14.857 "claimed": false, 00:11:14.857 "zoned": false, 00:11:14.857 "supported_io_types": { 00:11:14.857 "read": true, 00:11:14.857 "write": true, 00:11:14.857 "unmap": true, 00:11:14.857 "flush": true, 00:11:14.857 "reset": true, 00:11:14.857 "nvme_admin": false, 00:11:14.857 "nvme_io": false, 00:11:14.857 "nvme_io_md": false, 00:11:14.857 "write_zeroes": true, 00:11:14.857 "zcopy": true, 00:11:14.857 "get_zone_info": false, 00:11:14.857 "zone_management": false, 00:11:14.857 "zone_append": false, 00:11:14.857 "compare": false, 00:11:14.857 "compare_and_write": false, 00:11:14.857 "abort": true, 00:11:14.857 "seek_hole": false, 00:11:14.857 "seek_data": false, 00:11:14.857 "copy": true, 00:11:14.857 "nvme_iov_md": false 00:11:14.857 }, 00:11:14.857 "memory_domains": [ 00:11:14.857 { 00:11:14.857 "dma_device_id": "system", 00:11:14.857 "dma_device_type": 1 00:11:14.857 }, 00:11:14.857 { 00:11:14.857 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.857 "dma_device_type": 2 00:11:14.857 } 00:11:14.857 ], 00:11:14.857 "driver_specific": {} 00:11:14.857 } 00:11:14.857 ] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.857 BaseBdev4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.857 [ 00:11:14.857 { 00:11:14.857 "name": "BaseBdev4", 00:11:14.857 "aliases": [ 00:11:14.857 "f54a8383-50e8-4a53-919f-e1b137c2cef8" 00:11:14.857 ], 00:11:14.857 "product_name": "Malloc disk", 00:11:14.857 "block_size": 512, 00:11:14.857 "num_blocks": 65536, 00:11:14.857 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:14.857 "assigned_rate_limits": { 00:11:14.857 "rw_ios_per_sec": 0, 00:11:14.857 "rw_mbytes_per_sec": 0, 00:11:14.857 "r_mbytes_per_sec": 0, 00:11:14.857 "w_mbytes_per_sec": 0 00:11:14.857 }, 00:11:14.857 "claimed": false, 00:11:14.857 "zoned": false, 00:11:14.857 "supported_io_types": { 00:11:14.857 "read": true, 00:11:14.857 "write": true, 00:11:14.857 "unmap": true, 00:11:14.857 "flush": true, 00:11:14.857 "reset": true, 00:11:14.857 "nvme_admin": false, 00:11:14.857 "nvme_io": false, 00:11:14.857 "nvme_io_md": false, 00:11:14.857 "write_zeroes": true, 00:11:14.857 "zcopy": true, 00:11:14.857 "get_zone_info": false, 00:11:14.857 "zone_management": false, 00:11:14.857 "zone_append": false, 00:11:14.857 "compare": false, 00:11:14.857 "compare_and_write": false, 00:11:14.857 "abort": true, 00:11:14.857 "seek_hole": false, 00:11:14.857 "seek_data": false, 00:11:14.857 "copy": true, 00:11:14.857 "nvme_iov_md": false 00:11:14.857 }, 00:11:14.857 "memory_domains": [ 00:11:14.857 { 00:11:14.857 "dma_device_id": "system", 00:11:14.857 "dma_device_type": 1 00:11:14.857 }, 00:11:14.857 { 00:11:14.857 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.857 "dma_device_type": 2 00:11:14.857 } 00:11:14.857 ], 00:11:14.857 "driver_specific": {} 00:11:14.857 } 00:11:14.857 ] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.857 [2024-10-29 11:00:20.192560] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:14.857 [2024-10-29 11:00:20.192619] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:14.857 [2024-10-29 11:00:20.192644] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:14.857 [2024-10-29 11:00:20.194830] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:14.857 [2024-10-29 11:00:20.194881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.857 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.858 "name": "Existed_Raid", 00:11:14.858 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:14.858 "strip_size_kb": 0, 00:11:14.858 "state": "configuring", 00:11:14.858 "raid_level": "raid1", 00:11:14.858 "superblock": true, 00:11:14.858 "num_base_bdevs": 4, 00:11:14.858 "num_base_bdevs_discovered": 3, 00:11:14.858 "num_base_bdevs_operational": 4, 00:11:14.858 "base_bdevs_list": [ 00:11:14.858 { 00:11:14.858 "name": "BaseBdev1", 00:11:14.858 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.858 "is_configured": false, 00:11:14.858 "data_offset": 0, 00:11:14.858 "data_size": 0 00:11:14.858 }, 00:11:14.858 { 00:11:14.858 "name": "BaseBdev2", 00:11:14.858 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:14.858 "is_configured": true, 00:11:14.858 "data_offset": 2048, 00:11:14.858 "data_size": 63488 00:11:14.858 }, 00:11:14.858 { 00:11:14.858 "name": "BaseBdev3", 00:11:14.858 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:14.858 "is_configured": true, 00:11:14.858 "data_offset": 2048, 00:11:14.858 "data_size": 63488 00:11:14.858 }, 00:11:14.858 { 00:11:14.858 "name": "BaseBdev4", 00:11:14.858 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:14.858 "is_configured": true, 00:11:14.858 "data_offset": 2048, 00:11:14.858 "data_size": 63488 00:11:14.858 } 00:11:14.858 ] 00:11:14.858 }' 00:11:14.858 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.858 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.118 [2024-10-29 11:00:20.599912] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.118 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.378 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.378 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.378 "name": "Existed_Raid", 00:11:15.378 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:15.378 "strip_size_kb": 0, 00:11:15.378 "state": "configuring", 00:11:15.378 "raid_level": "raid1", 00:11:15.378 "superblock": true, 00:11:15.378 "num_base_bdevs": 4, 00:11:15.378 "num_base_bdevs_discovered": 2, 00:11:15.378 "num_base_bdevs_operational": 4, 00:11:15.378 "base_bdevs_list": [ 00:11:15.378 { 00:11:15.378 "name": "BaseBdev1", 00:11:15.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:15.378 "is_configured": false, 00:11:15.378 "data_offset": 0, 00:11:15.378 "data_size": 0 00:11:15.378 }, 00:11:15.378 { 00:11:15.378 "name": null, 00:11:15.378 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:15.378 "is_configured": false, 00:11:15.378 "data_offset": 0, 00:11:15.378 "data_size": 63488 00:11:15.378 }, 00:11:15.378 { 00:11:15.378 "name": "BaseBdev3", 00:11:15.378 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:15.378 "is_configured": true, 00:11:15.378 "data_offset": 2048, 00:11:15.378 "data_size": 63488 00:11:15.378 }, 00:11:15.378 { 00:11:15.378 "name": "BaseBdev4", 00:11:15.378 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:15.378 "is_configured": true, 00:11:15.378 "data_offset": 2048, 00:11:15.378 "data_size": 63488 00:11:15.378 } 00:11:15.378 ] 00:11:15.378 }' 00:11:15.378 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.378 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.637 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.637 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.637 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.637 11:00:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:15.637 11:00:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.637 [2024-10-29 11:00:21.044334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:15.637 BaseBdev1 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.637 [ 00:11:15.637 { 00:11:15.637 "name": "BaseBdev1", 00:11:15.637 "aliases": [ 00:11:15.637 "3ebb2cc2-cb24-42f4-8906-23628290d941" 00:11:15.637 ], 00:11:15.637 "product_name": "Malloc disk", 00:11:15.637 "block_size": 512, 00:11:15.637 "num_blocks": 65536, 00:11:15.637 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:15.637 "assigned_rate_limits": { 00:11:15.637 "rw_ios_per_sec": 0, 00:11:15.637 "rw_mbytes_per_sec": 0, 00:11:15.637 "r_mbytes_per_sec": 0, 00:11:15.637 "w_mbytes_per_sec": 0 00:11:15.637 }, 00:11:15.637 "claimed": true, 00:11:15.637 "claim_type": "exclusive_write", 00:11:15.637 "zoned": false, 00:11:15.637 "supported_io_types": { 00:11:15.637 "read": true, 00:11:15.637 "write": true, 00:11:15.637 "unmap": true, 00:11:15.637 "flush": true, 00:11:15.637 "reset": true, 00:11:15.637 "nvme_admin": false, 00:11:15.637 "nvme_io": false, 00:11:15.637 "nvme_io_md": false, 00:11:15.637 "write_zeroes": true, 00:11:15.637 "zcopy": true, 00:11:15.637 "get_zone_info": false, 00:11:15.637 "zone_management": false, 00:11:15.637 "zone_append": false, 00:11:15.637 "compare": false, 00:11:15.637 "compare_and_write": false, 00:11:15.637 "abort": true, 00:11:15.637 "seek_hole": false, 00:11:15.637 "seek_data": false, 00:11:15.637 "copy": true, 00:11:15.637 "nvme_iov_md": false 00:11:15.637 }, 00:11:15.637 "memory_domains": [ 00:11:15.637 { 00:11:15.637 "dma_device_id": "system", 00:11:15.637 "dma_device_type": 1 00:11:15.637 }, 00:11:15.637 { 00:11:15.637 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:15.637 "dma_device_type": 2 00:11:15.637 } 00:11:15.637 ], 00:11:15.637 "driver_specific": {} 00:11:15.637 } 00:11:15.637 ] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.637 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.638 "name": "Existed_Raid", 00:11:15.638 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:15.638 "strip_size_kb": 0, 00:11:15.638 "state": "configuring", 00:11:15.638 "raid_level": "raid1", 00:11:15.638 "superblock": true, 00:11:15.638 "num_base_bdevs": 4, 00:11:15.638 "num_base_bdevs_discovered": 3, 00:11:15.638 "num_base_bdevs_operational": 4, 00:11:15.638 "base_bdevs_list": [ 00:11:15.638 { 00:11:15.638 "name": "BaseBdev1", 00:11:15.638 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:15.638 "is_configured": true, 00:11:15.638 "data_offset": 2048, 00:11:15.638 "data_size": 63488 00:11:15.638 }, 00:11:15.638 { 00:11:15.638 "name": null, 00:11:15.638 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:15.638 "is_configured": false, 00:11:15.638 "data_offset": 0, 00:11:15.638 "data_size": 63488 00:11:15.638 }, 00:11:15.638 { 00:11:15.638 "name": "BaseBdev3", 00:11:15.638 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:15.638 "is_configured": true, 00:11:15.638 "data_offset": 2048, 00:11:15.638 "data_size": 63488 00:11:15.638 }, 00:11:15.638 { 00:11:15.638 "name": "BaseBdev4", 00:11:15.638 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:15.638 "is_configured": true, 00:11:15.638 "data_offset": 2048, 00:11:15.638 "data_size": 63488 00:11:15.638 } 00:11:15.638 ] 00:11:15.638 }' 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.638 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.207 [2024-10-29 11:00:21.583620] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.207 "name": "Existed_Raid", 00:11:16.207 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:16.207 "strip_size_kb": 0, 00:11:16.207 "state": "configuring", 00:11:16.207 "raid_level": "raid1", 00:11:16.207 "superblock": true, 00:11:16.207 "num_base_bdevs": 4, 00:11:16.207 "num_base_bdevs_discovered": 2, 00:11:16.207 "num_base_bdevs_operational": 4, 00:11:16.207 "base_bdevs_list": [ 00:11:16.207 { 00:11:16.207 "name": "BaseBdev1", 00:11:16.207 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:16.207 "is_configured": true, 00:11:16.207 "data_offset": 2048, 00:11:16.207 "data_size": 63488 00:11:16.207 }, 00:11:16.207 { 00:11:16.207 "name": null, 00:11:16.207 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:16.207 "is_configured": false, 00:11:16.207 "data_offset": 0, 00:11:16.207 "data_size": 63488 00:11:16.207 }, 00:11:16.207 { 00:11:16.207 "name": null, 00:11:16.207 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:16.207 "is_configured": false, 00:11:16.207 "data_offset": 0, 00:11:16.207 "data_size": 63488 00:11:16.207 }, 00:11:16.207 { 00:11:16.207 "name": "BaseBdev4", 00:11:16.207 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:16.207 "is_configured": true, 00:11:16.207 "data_offset": 2048, 00:11:16.207 "data_size": 63488 00:11:16.207 } 00:11:16.207 ] 00:11:16.207 }' 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.207 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.778 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.778 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.778 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.778 11:00:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:16.778 11:00:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.778 [2024-10-29 11:00:22.015329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.778 "name": "Existed_Raid", 00:11:16.778 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:16.778 "strip_size_kb": 0, 00:11:16.778 "state": "configuring", 00:11:16.778 "raid_level": "raid1", 00:11:16.778 "superblock": true, 00:11:16.778 "num_base_bdevs": 4, 00:11:16.778 "num_base_bdevs_discovered": 3, 00:11:16.778 "num_base_bdevs_operational": 4, 00:11:16.778 "base_bdevs_list": [ 00:11:16.778 { 00:11:16.778 "name": "BaseBdev1", 00:11:16.778 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:16.778 "is_configured": true, 00:11:16.778 "data_offset": 2048, 00:11:16.778 "data_size": 63488 00:11:16.778 }, 00:11:16.778 { 00:11:16.778 "name": null, 00:11:16.778 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:16.778 "is_configured": false, 00:11:16.778 "data_offset": 0, 00:11:16.778 "data_size": 63488 00:11:16.778 }, 00:11:16.778 { 00:11:16.778 "name": "BaseBdev3", 00:11:16.778 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:16.778 "is_configured": true, 00:11:16.778 "data_offset": 2048, 00:11:16.778 "data_size": 63488 00:11:16.778 }, 00:11:16.778 { 00:11:16.778 "name": "BaseBdev4", 00:11:16.778 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:16.778 "is_configured": true, 00:11:16.778 "data_offset": 2048, 00:11:16.778 "data_size": 63488 00:11:16.778 } 00:11:16.778 ] 00:11:16.778 }' 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.778 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.037 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:17.037 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.037 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.037 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.038 [2024-10-29 11:00:22.482495] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.038 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.297 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.297 "name": "Existed_Raid", 00:11:17.297 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:17.297 "strip_size_kb": 0, 00:11:17.297 "state": "configuring", 00:11:17.297 "raid_level": "raid1", 00:11:17.297 "superblock": true, 00:11:17.297 "num_base_bdevs": 4, 00:11:17.297 "num_base_bdevs_discovered": 2, 00:11:17.297 "num_base_bdevs_operational": 4, 00:11:17.297 "base_bdevs_list": [ 00:11:17.297 { 00:11:17.297 "name": null, 00:11:17.297 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:17.297 "is_configured": false, 00:11:17.297 "data_offset": 0, 00:11:17.297 "data_size": 63488 00:11:17.297 }, 00:11:17.297 { 00:11:17.297 "name": null, 00:11:17.297 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:17.297 "is_configured": false, 00:11:17.297 "data_offset": 0, 00:11:17.297 "data_size": 63488 00:11:17.297 }, 00:11:17.297 { 00:11:17.297 "name": "BaseBdev3", 00:11:17.297 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:17.297 "is_configured": true, 00:11:17.297 "data_offset": 2048, 00:11:17.297 "data_size": 63488 00:11:17.297 }, 00:11:17.297 { 00:11:17.297 "name": "BaseBdev4", 00:11:17.297 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:17.297 "is_configured": true, 00:11:17.297 "data_offset": 2048, 00:11:17.297 "data_size": 63488 00:11:17.297 } 00:11:17.297 ] 00:11:17.297 }' 00:11:17.297 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.297 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.558 [2024-10-29 11:00:22.988783] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.558 11:00:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.558 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.558 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.558 "name": "Existed_Raid", 00:11:17.558 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:17.558 "strip_size_kb": 0, 00:11:17.558 "state": "configuring", 00:11:17.558 "raid_level": "raid1", 00:11:17.558 "superblock": true, 00:11:17.558 "num_base_bdevs": 4, 00:11:17.558 "num_base_bdevs_discovered": 3, 00:11:17.558 "num_base_bdevs_operational": 4, 00:11:17.558 "base_bdevs_list": [ 00:11:17.558 { 00:11:17.558 "name": null, 00:11:17.558 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:17.558 "is_configured": false, 00:11:17.558 "data_offset": 0, 00:11:17.558 "data_size": 63488 00:11:17.558 }, 00:11:17.558 { 00:11:17.558 "name": "BaseBdev2", 00:11:17.558 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:17.558 "is_configured": true, 00:11:17.558 "data_offset": 2048, 00:11:17.558 "data_size": 63488 00:11:17.558 }, 00:11:17.558 { 00:11:17.558 "name": "BaseBdev3", 00:11:17.558 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:17.558 "is_configured": true, 00:11:17.558 "data_offset": 2048, 00:11:17.558 "data_size": 63488 00:11:17.558 }, 00:11:17.558 { 00:11:17.558 "name": "BaseBdev4", 00:11:17.558 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:17.558 "is_configured": true, 00:11:17.558 "data_offset": 2048, 00:11:17.558 "data_size": 63488 00:11:17.558 } 00:11:17.558 ] 00:11:17.558 }' 00:11:17.558 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.558 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3ebb2cc2-cb24-42f4-8906-23628290d941 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 [2024-10-29 11:00:23.557045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:18.129 [2024-10-29 11:00:23.557282] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:11:18.129 [2024-10-29 11:00:23.557307] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:18.129 NewBaseBdev 00:11:18.129 [2024-10-29 11:00:23.557683] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:11:18.129 [2024-10-29 11:00:23.557826] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:11:18.129 [2024-10-29 11:00:23.557838] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:11:18.129 [2024-10-29 11:00:23.557949] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 [ 00:11:18.129 { 00:11:18.129 "name": "NewBaseBdev", 00:11:18.129 "aliases": [ 00:11:18.129 "3ebb2cc2-cb24-42f4-8906-23628290d941" 00:11:18.129 ], 00:11:18.129 "product_name": "Malloc disk", 00:11:18.129 "block_size": 512, 00:11:18.129 "num_blocks": 65536, 00:11:18.129 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:18.129 "assigned_rate_limits": { 00:11:18.129 "rw_ios_per_sec": 0, 00:11:18.129 "rw_mbytes_per_sec": 0, 00:11:18.129 "r_mbytes_per_sec": 0, 00:11:18.129 "w_mbytes_per_sec": 0 00:11:18.129 }, 00:11:18.129 "claimed": true, 00:11:18.129 "claim_type": "exclusive_write", 00:11:18.129 "zoned": false, 00:11:18.129 "supported_io_types": { 00:11:18.129 "read": true, 00:11:18.129 "write": true, 00:11:18.129 "unmap": true, 00:11:18.129 "flush": true, 00:11:18.129 "reset": true, 00:11:18.129 "nvme_admin": false, 00:11:18.129 "nvme_io": false, 00:11:18.129 "nvme_io_md": false, 00:11:18.129 "write_zeroes": true, 00:11:18.129 "zcopy": true, 00:11:18.129 "get_zone_info": false, 00:11:18.129 "zone_management": false, 00:11:18.129 "zone_append": false, 00:11:18.129 "compare": false, 00:11:18.129 "compare_and_write": false, 00:11:18.129 "abort": true, 00:11:18.129 "seek_hole": false, 00:11:18.129 "seek_data": false, 00:11:18.129 "copy": true, 00:11:18.129 "nvme_iov_md": false 00:11:18.129 }, 00:11:18.129 "memory_domains": [ 00:11:18.129 { 00:11:18.129 "dma_device_id": "system", 00:11:18.129 "dma_device_type": 1 00:11:18.129 }, 00:11:18.129 { 00:11:18.129 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.129 "dma_device_type": 2 00:11:18.129 } 00:11:18.129 ], 00:11:18.129 "driver_specific": {} 00:11:18.129 } 00:11:18.129 ] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.129 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.390 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:18.390 "name": "Existed_Raid", 00:11:18.390 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:18.390 "strip_size_kb": 0, 00:11:18.390 "state": "online", 00:11:18.390 "raid_level": "raid1", 00:11:18.390 "superblock": true, 00:11:18.390 "num_base_bdevs": 4, 00:11:18.390 "num_base_bdevs_discovered": 4, 00:11:18.390 "num_base_bdevs_operational": 4, 00:11:18.390 "base_bdevs_list": [ 00:11:18.390 { 00:11:18.390 "name": "NewBaseBdev", 00:11:18.390 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:18.390 "is_configured": true, 00:11:18.390 "data_offset": 2048, 00:11:18.390 "data_size": 63488 00:11:18.390 }, 00:11:18.390 { 00:11:18.390 "name": "BaseBdev2", 00:11:18.390 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:18.390 "is_configured": true, 00:11:18.390 "data_offset": 2048, 00:11:18.390 "data_size": 63488 00:11:18.390 }, 00:11:18.390 { 00:11:18.390 "name": "BaseBdev3", 00:11:18.390 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:18.390 "is_configured": true, 00:11:18.390 "data_offset": 2048, 00:11:18.390 "data_size": 63488 00:11:18.390 }, 00:11:18.390 { 00:11:18.390 "name": "BaseBdev4", 00:11:18.390 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:18.390 "is_configured": true, 00:11:18.390 "data_offset": 2048, 00:11:18.390 "data_size": 63488 00:11:18.390 } 00:11:18.390 ] 00:11:18.390 }' 00:11:18.390 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:18.390 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.650 11:00:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.650 [2024-10-29 11:00:24.004862] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:18.650 "name": "Existed_Raid", 00:11:18.650 "aliases": [ 00:11:18.650 "c3cda496-c7da-4f48-8dc4-0aea0ecbf053" 00:11:18.650 ], 00:11:18.650 "product_name": "Raid Volume", 00:11:18.650 "block_size": 512, 00:11:18.650 "num_blocks": 63488, 00:11:18.650 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:18.650 "assigned_rate_limits": { 00:11:18.650 "rw_ios_per_sec": 0, 00:11:18.650 "rw_mbytes_per_sec": 0, 00:11:18.650 "r_mbytes_per_sec": 0, 00:11:18.650 "w_mbytes_per_sec": 0 00:11:18.650 }, 00:11:18.650 "claimed": false, 00:11:18.650 "zoned": false, 00:11:18.650 "supported_io_types": { 00:11:18.650 "read": true, 00:11:18.650 "write": true, 00:11:18.650 "unmap": false, 00:11:18.650 "flush": false, 00:11:18.650 "reset": true, 00:11:18.650 "nvme_admin": false, 00:11:18.650 "nvme_io": false, 00:11:18.650 "nvme_io_md": false, 00:11:18.650 "write_zeroes": true, 00:11:18.650 "zcopy": false, 00:11:18.650 "get_zone_info": false, 00:11:18.650 "zone_management": false, 00:11:18.650 "zone_append": false, 00:11:18.650 "compare": false, 00:11:18.650 "compare_and_write": false, 00:11:18.650 "abort": false, 00:11:18.650 "seek_hole": false, 00:11:18.650 "seek_data": false, 00:11:18.650 "copy": false, 00:11:18.650 "nvme_iov_md": false 00:11:18.650 }, 00:11:18.650 "memory_domains": [ 00:11:18.650 { 00:11:18.650 "dma_device_id": "system", 00:11:18.650 "dma_device_type": 1 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.650 "dma_device_type": 2 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "system", 00:11:18.650 "dma_device_type": 1 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.650 "dma_device_type": 2 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "system", 00:11:18.650 "dma_device_type": 1 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.650 "dma_device_type": 2 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "system", 00:11:18.650 "dma_device_type": 1 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.650 "dma_device_type": 2 00:11:18.650 } 00:11:18.650 ], 00:11:18.650 "driver_specific": { 00:11:18.650 "raid": { 00:11:18.650 "uuid": "c3cda496-c7da-4f48-8dc4-0aea0ecbf053", 00:11:18.650 "strip_size_kb": 0, 00:11:18.650 "state": "online", 00:11:18.650 "raid_level": "raid1", 00:11:18.650 "superblock": true, 00:11:18.650 "num_base_bdevs": 4, 00:11:18.650 "num_base_bdevs_discovered": 4, 00:11:18.650 "num_base_bdevs_operational": 4, 00:11:18.650 "base_bdevs_list": [ 00:11:18.650 { 00:11:18.650 "name": "NewBaseBdev", 00:11:18.650 "uuid": "3ebb2cc2-cb24-42f4-8906-23628290d941", 00:11:18.650 "is_configured": true, 00:11:18.650 "data_offset": 2048, 00:11:18.650 "data_size": 63488 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "name": "BaseBdev2", 00:11:18.650 "uuid": "afa6577d-5918-4b99-84bb-8c66400cfd33", 00:11:18.650 "is_configured": true, 00:11:18.650 "data_offset": 2048, 00:11:18.650 "data_size": 63488 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "name": "BaseBdev3", 00:11:18.650 "uuid": "f4e84ee4-88c4-4cfa-9933-b41071172319", 00:11:18.650 "is_configured": true, 00:11:18.650 "data_offset": 2048, 00:11:18.650 "data_size": 63488 00:11:18.650 }, 00:11:18.650 { 00:11:18.650 "name": "BaseBdev4", 00:11:18.650 "uuid": "f54a8383-50e8-4a53-919f-e1b137c2cef8", 00:11:18.650 "is_configured": true, 00:11:18.650 "data_offset": 2048, 00:11:18.650 "data_size": 63488 00:11:18.650 } 00:11:18.650 ] 00:11:18.650 } 00:11:18.650 } 00:11:18.650 }' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:18.650 BaseBdev2 00:11:18.650 BaseBdev3 00:11:18.650 BaseBdev4' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.650 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.911 [2024-10-29 11:00:24.299888] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:18.911 [2024-10-29 11:00:24.300009] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:18.911 [2024-10-29 11:00:24.300113] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:18.911 [2024-10-29 11:00:24.300439] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:18.911 [2024-10-29 11:00:24.300459] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84790 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 84790 ']' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 84790 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 84790 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 84790' 00:11:18.911 killing process with pid 84790 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 84790 00:11:18.911 [2024-10-29 11:00:24.346169] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:18.911 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 84790 00:11:19.171 [2024-10-29 11:00:24.425625] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:19.431 11:00:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:19.431 00:11:19.431 real 0m9.425s 00:11:19.431 user 0m15.811s 00:11:19.431 sys 0m2.078s 00:11:19.431 ************************************ 00:11:19.431 END TEST raid_state_function_test_sb 00:11:19.431 ************************************ 00:11:19.431 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:19.431 11:00:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.431 11:00:24 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:11:19.431 11:00:24 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:11:19.431 11:00:24 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:19.431 11:00:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:19.431 ************************************ 00:11:19.431 START TEST raid_superblock_test 00:11:19.431 ************************************ 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 4 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=85438 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 85438 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 85438 ']' 00:11:19.431 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:19.431 11:00:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.431 [2024-10-29 11:00:24.847035] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:19.431 [2024-10-29 11:00:24.847196] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85438 ] 00:11:19.691 [2024-10-29 11:00:25.006752] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:19.691 [2024-10-29 11:00:25.035291] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:19.691 [2024-10-29 11:00:25.081045] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:19.691 [2024-10-29 11:00:25.081088] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.664 malloc1 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.664 [2024-10-29 11:00:25.817189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:20.664 [2024-10-29 11:00:25.817326] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.664 [2024-10-29 11:00:25.817423] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:20.664 [2024-10-29 11:00:25.817479] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.664 [2024-10-29 11:00:25.820035] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.664 [2024-10-29 11:00:25.820123] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:20.664 pt1 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.664 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 malloc2 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 [2024-10-29 11:00:25.850735] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:20.665 [2024-10-29 11:00:25.850796] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.665 [2024-10-29 11:00:25.850815] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:20.665 [2024-10-29 11:00:25.850826] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.665 [2024-10-29 11:00:25.853249] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.665 [2024-10-29 11:00:25.853296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:20.665 pt2 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 malloc3 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 [2024-10-29 11:00:25.884158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:20.665 [2024-10-29 11:00:25.884210] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.665 [2024-10-29 11:00:25.884246] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:20.665 [2024-10-29 11:00:25.884258] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.665 [2024-10-29 11:00:25.886714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.665 [2024-10-29 11:00:25.886751] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:20.665 pt3 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 malloc4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 [2024-10-29 11:00:25.923459] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:20.665 [2024-10-29 11:00:25.923531] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.665 [2024-10-29 11:00:25.923552] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:20.665 [2024-10-29 11:00:25.923568] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.665 [2024-10-29 11:00:25.926284] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.665 [2024-10-29 11:00:25.926329] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:20.665 pt4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 [2024-10-29 11:00:25.935511] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:20.665 [2024-10-29 11:00:25.937655] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:20.665 [2024-10-29 11:00:25.937726] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:20.665 [2024-10-29 11:00:25.937776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:20.665 [2024-10-29 11:00:25.937974] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:20.665 [2024-10-29 11:00:25.938009] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:20.665 [2024-10-29 11:00:25.938348] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:11:20.665 [2024-10-29 11:00:25.938561] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:20.665 [2024-10-29 11:00:25.938588] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:20.665 [2024-10-29 11:00:25.938746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.665 "name": "raid_bdev1", 00:11:20.665 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:20.665 "strip_size_kb": 0, 00:11:20.665 "state": "online", 00:11:20.665 "raid_level": "raid1", 00:11:20.665 "superblock": true, 00:11:20.665 "num_base_bdevs": 4, 00:11:20.665 "num_base_bdevs_discovered": 4, 00:11:20.665 "num_base_bdevs_operational": 4, 00:11:20.665 "base_bdevs_list": [ 00:11:20.665 { 00:11:20.665 "name": "pt1", 00:11:20.665 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:20.665 "is_configured": true, 00:11:20.665 "data_offset": 2048, 00:11:20.665 "data_size": 63488 00:11:20.665 }, 00:11:20.665 { 00:11:20.665 "name": "pt2", 00:11:20.665 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:20.665 "is_configured": true, 00:11:20.665 "data_offset": 2048, 00:11:20.665 "data_size": 63488 00:11:20.665 }, 00:11:20.665 { 00:11:20.665 "name": "pt3", 00:11:20.665 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:20.665 "is_configured": true, 00:11:20.665 "data_offset": 2048, 00:11:20.665 "data_size": 63488 00:11:20.665 }, 00:11:20.665 { 00:11:20.665 "name": "pt4", 00:11:20.665 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:20.665 "is_configured": true, 00:11:20.665 "data_offset": 2048, 00:11:20.665 "data_size": 63488 00:11:20.665 } 00:11:20.665 ] 00:11:20.665 }' 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.665 11:00:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.236 [2024-10-29 11:00:26.439045] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:21.236 "name": "raid_bdev1", 00:11:21.236 "aliases": [ 00:11:21.236 "f9597dbe-de74-488f-a9c1-35d4897f0c35" 00:11:21.236 ], 00:11:21.236 "product_name": "Raid Volume", 00:11:21.236 "block_size": 512, 00:11:21.236 "num_blocks": 63488, 00:11:21.236 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:21.236 "assigned_rate_limits": { 00:11:21.236 "rw_ios_per_sec": 0, 00:11:21.236 "rw_mbytes_per_sec": 0, 00:11:21.236 "r_mbytes_per_sec": 0, 00:11:21.236 "w_mbytes_per_sec": 0 00:11:21.236 }, 00:11:21.236 "claimed": false, 00:11:21.236 "zoned": false, 00:11:21.236 "supported_io_types": { 00:11:21.236 "read": true, 00:11:21.236 "write": true, 00:11:21.236 "unmap": false, 00:11:21.236 "flush": false, 00:11:21.236 "reset": true, 00:11:21.236 "nvme_admin": false, 00:11:21.236 "nvme_io": false, 00:11:21.236 "nvme_io_md": false, 00:11:21.236 "write_zeroes": true, 00:11:21.236 "zcopy": false, 00:11:21.236 "get_zone_info": false, 00:11:21.236 "zone_management": false, 00:11:21.236 "zone_append": false, 00:11:21.236 "compare": false, 00:11:21.236 "compare_and_write": false, 00:11:21.236 "abort": false, 00:11:21.236 "seek_hole": false, 00:11:21.236 "seek_data": false, 00:11:21.236 "copy": false, 00:11:21.236 "nvme_iov_md": false 00:11:21.236 }, 00:11:21.236 "memory_domains": [ 00:11:21.236 { 00:11:21.236 "dma_device_id": "system", 00:11:21.236 "dma_device_type": 1 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.236 "dma_device_type": 2 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "system", 00:11:21.236 "dma_device_type": 1 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.236 "dma_device_type": 2 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "system", 00:11:21.236 "dma_device_type": 1 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.236 "dma_device_type": 2 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "system", 00:11:21.236 "dma_device_type": 1 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:21.236 "dma_device_type": 2 00:11:21.236 } 00:11:21.236 ], 00:11:21.236 "driver_specific": { 00:11:21.236 "raid": { 00:11:21.236 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:21.236 "strip_size_kb": 0, 00:11:21.236 "state": "online", 00:11:21.236 "raid_level": "raid1", 00:11:21.236 "superblock": true, 00:11:21.236 "num_base_bdevs": 4, 00:11:21.236 "num_base_bdevs_discovered": 4, 00:11:21.236 "num_base_bdevs_operational": 4, 00:11:21.236 "base_bdevs_list": [ 00:11:21.236 { 00:11:21.236 "name": "pt1", 00:11:21.236 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:21.236 "is_configured": true, 00:11:21.236 "data_offset": 2048, 00:11:21.236 "data_size": 63488 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "name": "pt2", 00:11:21.236 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:21.236 "is_configured": true, 00:11:21.236 "data_offset": 2048, 00:11:21.236 "data_size": 63488 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "name": "pt3", 00:11:21.236 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:21.236 "is_configured": true, 00:11:21.236 "data_offset": 2048, 00:11:21.236 "data_size": 63488 00:11:21.236 }, 00:11:21.236 { 00:11:21.236 "name": "pt4", 00:11:21.236 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:21.236 "is_configured": true, 00:11:21.236 "data_offset": 2048, 00:11:21.236 "data_size": 63488 00:11:21.236 } 00:11:21.236 ] 00:11:21.236 } 00:11:21.236 } 00:11:21.236 }' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:21.236 pt2 00:11:21.236 pt3 00:11:21.236 pt4' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.236 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.237 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 [2024-10-29 11:00:26.762515] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f9597dbe-de74-488f-a9c1-35d4897f0c35 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f9597dbe-de74-488f-a9c1-35d4897f0c35 ']' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 [2024-10-29 11:00:26.810025] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:21.498 [2024-10-29 11:00:26.810064] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:21.498 [2024-10-29 11:00:26.810150] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:21.498 [2024-10-29 11:00:26.810251] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:21.498 [2024-10-29 11:00:26.810271] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.498 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.499 [2024-10-29 11:00:26.973800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:21.499 [2024-10-29 11:00:26.975898] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:21.499 [2024-10-29 11:00:26.975965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:21.499 [2024-10-29 11:00:26.976001] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:21.499 [2024-10-29 11:00:26.976057] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:21.499 [2024-10-29 11:00:26.976106] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:21.499 [2024-10-29 11:00:26.976147] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:21.499 [2024-10-29 11:00:26.976168] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:21.499 [2024-10-29 11:00:26.976186] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:21.499 [2024-10-29 11:00:26.976197] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:11:21.499 request: 00:11:21.499 { 00:11:21.499 "name": "raid_bdev1", 00:11:21.499 "raid_level": "raid1", 00:11:21.499 "base_bdevs": [ 00:11:21.499 "malloc1", 00:11:21.499 "malloc2", 00:11:21.499 "malloc3", 00:11:21.499 "malloc4" 00:11:21.499 ], 00:11:21.499 "superblock": false, 00:11:21.499 "method": "bdev_raid_create", 00:11:21.499 "req_id": 1 00:11:21.499 } 00:11:21.499 Got JSON-RPC error response 00:11:21.499 response: 00:11:21.499 { 00:11:21.499 "code": -17, 00:11:21.499 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:21.499 } 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:21.499 11:00:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.761 [2024-10-29 11:00:27.041603] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:21.761 [2024-10-29 11:00:27.041660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:21.761 [2024-10-29 11:00:27.041684] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:21.761 [2024-10-29 11:00:27.041695] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:21.761 [2024-10-29 11:00:27.044191] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:21.761 [2024-10-29 11:00:27.044232] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:21.761 [2024-10-29 11:00:27.044313] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:21.761 [2024-10-29 11:00:27.044353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:21.761 pt1 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.761 "name": "raid_bdev1", 00:11:21.761 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:21.761 "strip_size_kb": 0, 00:11:21.761 "state": "configuring", 00:11:21.761 "raid_level": "raid1", 00:11:21.761 "superblock": true, 00:11:21.761 "num_base_bdevs": 4, 00:11:21.761 "num_base_bdevs_discovered": 1, 00:11:21.761 "num_base_bdevs_operational": 4, 00:11:21.761 "base_bdevs_list": [ 00:11:21.761 { 00:11:21.761 "name": "pt1", 00:11:21.761 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:21.761 "is_configured": true, 00:11:21.761 "data_offset": 2048, 00:11:21.761 "data_size": 63488 00:11:21.761 }, 00:11:21.761 { 00:11:21.761 "name": null, 00:11:21.761 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:21.761 "is_configured": false, 00:11:21.761 "data_offset": 2048, 00:11:21.761 "data_size": 63488 00:11:21.761 }, 00:11:21.761 { 00:11:21.761 "name": null, 00:11:21.761 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:21.761 "is_configured": false, 00:11:21.761 "data_offset": 2048, 00:11:21.761 "data_size": 63488 00:11:21.761 }, 00:11:21.761 { 00:11:21.761 "name": null, 00:11:21.761 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:21.761 "is_configured": false, 00:11:21.761 "data_offset": 2048, 00:11:21.761 "data_size": 63488 00:11:21.761 } 00:11:21.761 ] 00:11:21.761 }' 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.761 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.022 [2024-10-29 11:00:27.488879] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:22.022 [2024-10-29 11:00:27.488945] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.022 [2024-10-29 11:00:27.488966] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:22.022 [2024-10-29 11:00:27.488974] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.022 [2024-10-29 11:00:27.489429] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.022 [2024-10-29 11:00:27.489456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:22.022 [2024-10-29 11:00:27.489538] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:22.022 [2024-10-29 11:00:27.489595] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:22.022 pt2 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.022 [2024-10-29 11:00:27.500868] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.022 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.284 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.284 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.284 "name": "raid_bdev1", 00:11:22.284 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:22.284 "strip_size_kb": 0, 00:11:22.284 "state": "configuring", 00:11:22.284 "raid_level": "raid1", 00:11:22.284 "superblock": true, 00:11:22.284 "num_base_bdevs": 4, 00:11:22.284 "num_base_bdevs_discovered": 1, 00:11:22.284 "num_base_bdevs_operational": 4, 00:11:22.284 "base_bdevs_list": [ 00:11:22.284 { 00:11:22.284 "name": "pt1", 00:11:22.284 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:22.284 "is_configured": true, 00:11:22.284 "data_offset": 2048, 00:11:22.284 "data_size": 63488 00:11:22.284 }, 00:11:22.284 { 00:11:22.284 "name": null, 00:11:22.284 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:22.284 "is_configured": false, 00:11:22.284 "data_offset": 0, 00:11:22.284 "data_size": 63488 00:11:22.284 }, 00:11:22.284 { 00:11:22.284 "name": null, 00:11:22.284 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:22.284 "is_configured": false, 00:11:22.284 "data_offset": 2048, 00:11:22.284 "data_size": 63488 00:11:22.284 }, 00:11:22.284 { 00:11:22.284 "name": null, 00:11:22.284 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:22.284 "is_configured": false, 00:11:22.284 "data_offset": 2048, 00:11:22.284 "data_size": 63488 00:11:22.284 } 00:11:22.284 ] 00:11:22.284 }' 00:11:22.284 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.284 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.545 [2024-10-29 11:00:27.960089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:22.545 [2024-10-29 11:00:27.960167] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.545 [2024-10-29 11:00:27.960187] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:11:22.545 [2024-10-29 11:00:27.960200] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.545 [2024-10-29 11:00:27.960690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.545 [2024-10-29 11:00:27.960724] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:22.545 [2024-10-29 11:00:27.960817] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:22.545 [2024-10-29 11:00:27.960864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:22.545 pt2 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.545 [2024-10-29 11:00:27.972032] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:22.545 [2024-10-29 11:00:27.972091] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.545 [2024-10-29 11:00:27.972107] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:22.545 [2024-10-29 11:00:27.972119] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.545 [2024-10-29 11:00:27.972503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.545 [2024-10-29 11:00:27.972536] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:22.545 [2024-10-29 11:00:27.972599] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:22.545 [2024-10-29 11:00:27.972623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:22.545 pt3 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.545 [2024-10-29 11:00:27.983994] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:22.545 [2024-10-29 11:00:27.984053] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.545 [2024-10-29 11:00:27.984068] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:22.545 [2024-10-29 11:00:27.984079] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.545 [2024-10-29 11:00:27.984469] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.545 [2024-10-29 11:00:27.984500] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:22.545 [2024-10-29 11:00:27.984562] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:22.545 [2024-10-29 11:00:27.984585] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:22.545 [2024-10-29 11:00:27.984740] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:11:22.545 [2024-10-29 11:00:27.984764] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:22.545 [2024-10-29 11:00:27.985050] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:22.545 [2024-10-29 11:00:27.985221] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:11:22.545 [2024-10-29 11:00:27.985242] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:11:22.545 [2024-10-29 11:00:27.985385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:22.545 pt4 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.545 11:00:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.545 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.545 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.545 "name": "raid_bdev1", 00:11:22.545 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:22.545 "strip_size_kb": 0, 00:11:22.545 "state": "online", 00:11:22.545 "raid_level": "raid1", 00:11:22.545 "superblock": true, 00:11:22.545 "num_base_bdevs": 4, 00:11:22.545 "num_base_bdevs_discovered": 4, 00:11:22.545 "num_base_bdevs_operational": 4, 00:11:22.545 "base_bdevs_list": [ 00:11:22.545 { 00:11:22.545 "name": "pt1", 00:11:22.545 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:22.545 "is_configured": true, 00:11:22.545 "data_offset": 2048, 00:11:22.545 "data_size": 63488 00:11:22.545 }, 00:11:22.545 { 00:11:22.545 "name": "pt2", 00:11:22.545 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:22.545 "is_configured": true, 00:11:22.545 "data_offset": 2048, 00:11:22.545 "data_size": 63488 00:11:22.545 }, 00:11:22.545 { 00:11:22.545 "name": "pt3", 00:11:22.545 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:22.545 "is_configured": true, 00:11:22.545 "data_offset": 2048, 00:11:22.545 "data_size": 63488 00:11:22.545 }, 00:11:22.545 { 00:11:22.545 "name": "pt4", 00:11:22.545 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:22.545 "is_configured": true, 00:11:22.545 "data_offset": 2048, 00:11:22.545 "data_size": 63488 00:11:22.545 } 00:11:22.545 ] 00:11:22.545 }' 00:11:22.545 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.545 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:23.115 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.116 [2024-10-29 11:00:28.435683] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:23.116 "name": "raid_bdev1", 00:11:23.116 "aliases": [ 00:11:23.116 "f9597dbe-de74-488f-a9c1-35d4897f0c35" 00:11:23.116 ], 00:11:23.116 "product_name": "Raid Volume", 00:11:23.116 "block_size": 512, 00:11:23.116 "num_blocks": 63488, 00:11:23.116 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:23.116 "assigned_rate_limits": { 00:11:23.116 "rw_ios_per_sec": 0, 00:11:23.116 "rw_mbytes_per_sec": 0, 00:11:23.116 "r_mbytes_per_sec": 0, 00:11:23.116 "w_mbytes_per_sec": 0 00:11:23.116 }, 00:11:23.116 "claimed": false, 00:11:23.116 "zoned": false, 00:11:23.116 "supported_io_types": { 00:11:23.116 "read": true, 00:11:23.116 "write": true, 00:11:23.116 "unmap": false, 00:11:23.116 "flush": false, 00:11:23.116 "reset": true, 00:11:23.116 "nvme_admin": false, 00:11:23.116 "nvme_io": false, 00:11:23.116 "nvme_io_md": false, 00:11:23.116 "write_zeroes": true, 00:11:23.116 "zcopy": false, 00:11:23.116 "get_zone_info": false, 00:11:23.116 "zone_management": false, 00:11:23.116 "zone_append": false, 00:11:23.116 "compare": false, 00:11:23.116 "compare_and_write": false, 00:11:23.116 "abort": false, 00:11:23.116 "seek_hole": false, 00:11:23.116 "seek_data": false, 00:11:23.116 "copy": false, 00:11:23.116 "nvme_iov_md": false 00:11:23.116 }, 00:11:23.116 "memory_domains": [ 00:11:23.116 { 00:11:23.116 "dma_device_id": "system", 00:11:23.116 "dma_device_type": 1 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.116 "dma_device_type": 2 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "system", 00:11:23.116 "dma_device_type": 1 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.116 "dma_device_type": 2 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "system", 00:11:23.116 "dma_device_type": 1 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.116 "dma_device_type": 2 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "system", 00:11:23.116 "dma_device_type": 1 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:23.116 "dma_device_type": 2 00:11:23.116 } 00:11:23.116 ], 00:11:23.116 "driver_specific": { 00:11:23.116 "raid": { 00:11:23.116 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:23.116 "strip_size_kb": 0, 00:11:23.116 "state": "online", 00:11:23.116 "raid_level": "raid1", 00:11:23.116 "superblock": true, 00:11:23.116 "num_base_bdevs": 4, 00:11:23.116 "num_base_bdevs_discovered": 4, 00:11:23.116 "num_base_bdevs_operational": 4, 00:11:23.116 "base_bdevs_list": [ 00:11:23.116 { 00:11:23.116 "name": "pt1", 00:11:23.116 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:23.116 "is_configured": true, 00:11:23.116 "data_offset": 2048, 00:11:23.116 "data_size": 63488 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "name": "pt2", 00:11:23.116 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:23.116 "is_configured": true, 00:11:23.116 "data_offset": 2048, 00:11:23.116 "data_size": 63488 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "name": "pt3", 00:11:23.116 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:23.116 "is_configured": true, 00:11:23.116 "data_offset": 2048, 00:11:23.116 "data_size": 63488 00:11:23.116 }, 00:11:23.116 { 00:11:23.116 "name": "pt4", 00:11:23.116 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:23.116 "is_configured": true, 00:11:23.116 "data_offset": 2048, 00:11:23.116 "data_size": 63488 00:11:23.116 } 00:11:23.116 ] 00:11:23.116 } 00:11:23.116 } 00:11:23.116 }' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:23.116 pt2 00:11:23.116 pt3 00:11:23.116 pt4' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.116 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 [2024-10-29 11:00:28.751081] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f9597dbe-de74-488f-a9c1-35d4897f0c35 '!=' f9597dbe-de74-488f-a9c1-35d4897f0c35 ']' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 [2024-10-29 11:00:28.794691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.377 "name": "raid_bdev1", 00:11:23.377 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:23.377 "strip_size_kb": 0, 00:11:23.377 "state": "online", 00:11:23.377 "raid_level": "raid1", 00:11:23.377 "superblock": true, 00:11:23.377 "num_base_bdevs": 4, 00:11:23.377 "num_base_bdevs_discovered": 3, 00:11:23.377 "num_base_bdevs_operational": 3, 00:11:23.377 "base_bdevs_list": [ 00:11:23.377 { 00:11:23.377 "name": null, 00:11:23.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.377 "is_configured": false, 00:11:23.377 "data_offset": 0, 00:11:23.377 "data_size": 63488 00:11:23.377 }, 00:11:23.377 { 00:11:23.377 "name": "pt2", 00:11:23.377 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:23.377 "is_configured": true, 00:11:23.377 "data_offset": 2048, 00:11:23.377 "data_size": 63488 00:11:23.377 }, 00:11:23.377 { 00:11:23.377 "name": "pt3", 00:11:23.377 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:23.377 "is_configured": true, 00:11:23.377 "data_offset": 2048, 00:11:23.377 "data_size": 63488 00:11:23.377 }, 00:11:23.377 { 00:11:23.377 "name": "pt4", 00:11:23.377 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:23.377 "is_configured": true, 00:11:23.377 "data_offset": 2048, 00:11:23.377 "data_size": 63488 00:11:23.377 } 00:11:23.377 ] 00:11:23.377 }' 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.377 11:00:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 [2024-10-29 11:00:29.249942] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:23.948 [2024-10-29 11:00:29.249978] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:23.948 [2024-10-29 11:00:29.250126] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:23.948 [2024-10-29 11:00:29.250219] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:23.948 [2024-10-29 11:00:29.250275] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 [2024-10-29 11:00:29.349736] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:23.948 [2024-10-29 11:00:29.349799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:23.948 [2024-10-29 11:00:29.349818] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:11:23.948 [2024-10-29 11:00:29.349831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:23.948 [2024-10-29 11:00:29.352347] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:23.948 [2024-10-29 11:00:29.352406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:23.948 [2024-10-29 11:00:29.352482] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:23.948 [2024-10-29 11:00:29.352521] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:23.948 pt2 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.948 "name": "raid_bdev1", 00:11:23.948 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:23.948 "strip_size_kb": 0, 00:11:23.948 "state": "configuring", 00:11:23.948 "raid_level": "raid1", 00:11:23.948 "superblock": true, 00:11:23.948 "num_base_bdevs": 4, 00:11:23.948 "num_base_bdevs_discovered": 1, 00:11:23.948 "num_base_bdevs_operational": 3, 00:11:23.948 "base_bdevs_list": [ 00:11:23.948 { 00:11:23.948 "name": null, 00:11:23.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.948 "is_configured": false, 00:11:23.948 "data_offset": 2048, 00:11:23.948 "data_size": 63488 00:11:23.948 }, 00:11:23.948 { 00:11:23.948 "name": "pt2", 00:11:23.948 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:23.948 "is_configured": true, 00:11:23.948 "data_offset": 2048, 00:11:23.948 "data_size": 63488 00:11:23.948 }, 00:11:23.948 { 00:11:23.948 "name": null, 00:11:23.948 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:23.948 "is_configured": false, 00:11:23.948 "data_offset": 2048, 00:11:23.948 "data_size": 63488 00:11:23.948 }, 00:11:23.948 { 00:11:23.948 "name": null, 00:11:23.948 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:23.948 "is_configured": false, 00:11:23.948 "data_offset": 2048, 00:11:23.948 "data_size": 63488 00:11:23.948 } 00:11:23.948 ] 00:11:23.948 }' 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.948 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.518 [2024-10-29 11:00:29.800981] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:24.518 [2024-10-29 11:00:29.801049] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.518 [2024-10-29 11:00:29.801071] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:11:24.518 [2024-10-29 11:00:29.801084] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.518 [2024-10-29 11:00:29.801526] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.518 [2024-10-29 11:00:29.801559] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:24.518 [2024-10-29 11:00:29.801638] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:24.518 [2024-10-29 11:00:29.801676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:24.518 pt3 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.518 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.518 "name": "raid_bdev1", 00:11:24.518 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:24.518 "strip_size_kb": 0, 00:11:24.518 "state": "configuring", 00:11:24.518 "raid_level": "raid1", 00:11:24.518 "superblock": true, 00:11:24.518 "num_base_bdevs": 4, 00:11:24.518 "num_base_bdevs_discovered": 2, 00:11:24.518 "num_base_bdevs_operational": 3, 00:11:24.518 "base_bdevs_list": [ 00:11:24.518 { 00:11:24.518 "name": null, 00:11:24.518 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.518 "is_configured": false, 00:11:24.518 "data_offset": 2048, 00:11:24.518 "data_size": 63488 00:11:24.518 }, 00:11:24.518 { 00:11:24.518 "name": "pt2", 00:11:24.518 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:24.518 "is_configured": true, 00:11:24.518 "data_offset": 2048, 00:11:24.518 "data_size": 63488 00:11:24.518 }, 00:11:24.518 { 00:11:24.518 "name": "pt3", 00:11:24.518 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:24.518 "is_configured": true, 00:11:24.518 "data_offset": 2048, 00:11:24.518 "data_size": 63488 00:11:24.518 }, 00:11:24.518 { 00:11:24.519 "name": null, 00:11:24.519 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:24.519 "is_configured": false, 00:11:24.519 "data_offset": 2048, 00:11:24.519 "data_size": 63488 00:11:24.519 } 00:11:24.519 ] 00:11:24.519 }' 00:11:24.519 11:00:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.519 11:00:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.089 [2024-10-29 11:00:30.296151] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:25.089 [2024-10-29 11:00:30.296231] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:25.089 [2024-10-29 11:00:30.296256] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:11:25.089 [2024-10-29 11:00:30.296269] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:25.089 [2024-10-29 11:00:30.296756] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:25.089 [2024-10-29 11:00:30.296791] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:25.089 [2024-10-29 11:00:30.296872] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:25.089 [2024-10-29 11:00:30.296937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:25.089 [2024-10-29 11:00:30.297108] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:11:25.089 [2024-10-29 11:00:30.297131] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:25.089 [2024-10-29 11:00:30.297430] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:25.089 [2024-10-29 11:00:30.297598] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:11:25.089 [2024-10-29 11:00:30.297615] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:11:25.089 [2024-10-29 11:00:30.297754] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.089 pt4 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.089 "name": "raid_bdev1", 00:11:25.089 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:25.089 "strip_size_kb": 0, 00:11:25.089 "state": "online", 00:11:25.089 "raid_level": "raid1", 00:11:25.089 "superblock": true, 00:11:25.089 "num_base_bdevs": 4, 00:11:25.089 "num_base_bdevs_discovered": 3, 00:11:25.089 "num_base_bdevs_operational": 3, 00:11:25.089 "base_bdevs_list": [ 00:11:25.089 { 00:11:25.089 "name": null, 00:11:25.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.089 "is_configured": false, 00:11:25.089 "data_offset": 2048, 00:11:25.089 "data_size": 63488 00:11:25.089 }, 00:11:25.089 { 00:11:25.089 "name": "pt2", 00:11:25.089 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:25.089 "is_configured": true, 00:11:25.089 "data_offset": 2048, 00:11:25.089 "data_size": 63488 00:11:25.089 }, 00:11:25.089 { 00:11:25.089 "name": "pt3", 00:11:25.089 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:25.089 "is_configured": true, 00:11:25.089 "data_offset": 2048, 00:11:25.089 "data_size": 63488 00:11:25.089 }, 00:11:25.089 { 00:11:25.089 "name": "pt4", 00:11:25.089 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:25.089 "is_configured": true, 00:11:25.089 "data_offset": 2048, 00:11:25.089 "data_size": 63488 00:11:25.089 } 00:11:25.089 ] 00:11:25.089 }' 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.089 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 [2024-10-29 11:00:30.731530] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:25.349 [2024-10-29 11:00:30.731569] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:25.349 [2024-10-29 11:00:30.731668] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:25.349 [2024-10-29 11:00:30.731805] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:25.349 [2024-10-29 11:00:30.731825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 [2024-10-29 11:00:30.803403] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:25.349 [2024-10-29 11:00:30.803471] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:25.349 [2024-10-29 11:00:30.803510] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:11:25.349 [2024-10-29 11:00:30.803522] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:25.349 [2024-10-29 11:00:30.806158] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:25.349 [2024-10-29 11:00:30.806203] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:25.349 [2024-10-29 11:00:30.806288] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:25.349 [2024-10-29 11:00:30.806344] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:25.349 [2024-10-29 11:00:30.806555] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:11:25.349 [2024-10-29 11:00:30.806596] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:25.349 [2024-10-29 11:00:30.806619] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:11:25.349 [2024-10-29 11:00:30.806663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:25.349 [2024-10-29 11:00:30.806821] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:25.349 pt1 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.349 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.609 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.609 "name": "raid_bdev1", 00:11:25.609 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:25.609 "strip_size_kb": 0, 00:11:25.609 "state": "configuring", 00:11:25.609 "raid_level": "raid1", 00:11:25.609 "superblock": true, 00:11:25.609 "num_base_bdevs": 4, 00:11:25.609 "num_base_bdevs_discovered": 2, 00:11:25.609 "num_base_bdevs_operational": 3, 00:11:25.609 "base_bdevs_list": [ 00:11:25.609 { 00:11:25.609 "name": null, 00:11:25.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.609 "is_configured": false, 00:11:25.609 "data_offset": 2048, 00:11:25.609 "data_size": 63488 00:11:25.609 }, 00:11:25.609 { 00:11:25.609 "name": "pt2", 00:11:25.609 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:25.609 "is_configured": true, 00:11:25.609 "data_offset": 2048, 00:11:25.609 "data_size": 63488 00:11:25.609 }, 00:11:25.609 { 00:11:25.609 "name": "pt3", 00:11:25.609 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:25.609 "is_configured": true, 00:11:25.609 "data_offset": 2048, 00:11:25.609 "data_size": 63488 00:11:25.609 }, 00:11:25.609 { 00:11:25.609 "name": null, 00:11:25.609 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:25.609 "is_configured": false, 00:11:25.609 "data_offset": 2048, 00:11:25.609 "data_size": 63488 00:11:25.609 } 00:11:25.609 ] 00:11:25.609 }' 00:11:25.609 11:00:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.609 11:00:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.870 [2024-10-29 11:00:31.286576] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:25.870 [2024-10-29 11:00:31.286646] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:25.870 [2024-10-29 11:00:31.286670] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:11:25.870 [2024-10-29 11:00:31.286682] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:25.870 [2024-10-29 11:00:31.287180] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:25.870 [2024-10-29 11:00:31.287214] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:25.870 [2024-10-29 11:00:31.287307] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:25.870 [2024-10-29 11:00:31.287372] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:25.870 [2024-10-29 11:00:31.287516] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:11:25.870 [2024-10-29 11:00:31.287538] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:25.870 [2024-10-29 11:00:31.287841] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:11:25.870 [2024-10-29 11:00:31.288004] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:11:25.870 [2024-10-29 11:00:31.288025] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:11:25.870 [2024-10-29 11:00:31.288168] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.870 pt4 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.870 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.871 "name": "raid_bdev1", 00:11:25.871 "uuid": "f9597dbe-de74-488f-a9c1-35d4897f0c35", 00:11:25.871 "strip_size_kb": 0, 00:11:25.871 "state": "online", 00:11:25.871 "raid_level": "raid1", 00:11:25.871 "superblock": true, 00:11:25.871 "num_base_bdevs": 4, 00:11:25.871 "num_base_bdevs_discovered": 3, 00:11:25.871 "num_base_bdevs_operational": 3, 00:11:25.871 "base_bdevs_list": [ 00:11:25.871 { 00:11:25.871 "name": null, 00:11:25.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.871 "is_configured": false, 00:11:25.871 "data_offset": 2048, 00:11:25.871 "data_size": 63488 00:11:25.871 }, 00:11:25.871 { 00:11:25.871 "name": "pt2", 00:11:25.871 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:25.871 "is_configured": true, 00:11:25.871 "data_offset": 2048, 00:11:25.871 "data_size": 63488 00:11:25.871 }, 00:11:25.871 { 00:11:25.871 "name": "pt3", 00:11:25.871 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:25.871 "is_configured": true, 00:11:25.871 "data_offset": 2048, 00:11:25.871 "data_size": 63488 00:11:25.871 }, 00:11:25.871 { 00:11:25.871 "name": "pt4", 00:11:25.871 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:25.871 "is_configured": true, 00:11:25.871 "data_offset": 2048, 00:11:25.871 "data_size": 63488 00:11:25.871 } 00:11:25.871 ] 00:11:25.871 }' 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.871 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:11:26.441 [2024-10-29 11:00:31.770143] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' f9597dbe-de74-488f-a9c1-35d4897f0c35 '!=' f9597dbe-de74-488f-a9c1-35d4897f0c35 ']' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 85438 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 85438 ']' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # kill -0 85438 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # uname 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 85438 00:11:26.441 killing process with pid 85438 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 85438' 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@971 -- # kill 85438 00:11:26.441 [2024-10-29 11:00:31.851018] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:26.441 [2024-10-29 11:00:31.851115] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:26.441 11:00:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@976 -- # wait 85438 00:11:26.441 [2024-10-29 11:00:31.851204] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:26.441 [2024-10-29 11:00:31.851216] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:11:26.441 [2024-10-29 11:00:31.896874] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:26.701 11:00:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:26.701 00:11:26.701 real 0m7.372s 00:11:26.701 user 0m12.520s 00:11:26.701 sys 0m1.535s 00:11:26.701 11:00:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:26.701 11:00:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.701 ************************************ 00:11:26.701 END TEST raid_superblock_test 00:11:26.701 ************************************ 00:11:26.701 11:00:32 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:11:26.701 11:00:32 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:11:26.701 11:00:32 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:26.701 11:00:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:26.701 ************************************ 00:11:26.701 START TEST raid_read_error_test 00:11:26.701 ************************************ 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 4 read 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.701 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.X8GMJiOjKE 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85914 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85914 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@833 -- # '[' -z 85914 ']' 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:26.961 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:26.961 11:00:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.961 [2024-10-29 11:00:32.301644] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:26.961 [2024-10-29 11:00:32.301798] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85914 ] 00:11:27.221 [2024-10-29 11:00:32.477786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:27.221 [2024-10-29 11:00:32.504863] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:27.221 [2024-10-29 11:00:32.550763] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:27.221 [2024-10-29 11:00:32.550814] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@866 -- # return 0 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 BaseBdev1_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 true 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 [2024-10-29 11:00:33.215180] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:27.792 [2024-10-29 11:00:33.215238] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.792 [2024-10-29 11:00:33.215259] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:27.792 [2024-10-29 11:00:33.215270] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.792 [2024-10-29 11:00:33.217816] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.792 [2024-10-29 11:00:33.217860] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:27.792 BaseBdev1 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 BaseBdev2_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 true 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 [2024-10-29 11:00:33.256571] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:27.792 [2024-10-29 11:00:33.256630] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.792 [2024-10-29 11:00:33.256656] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:27.792 [2024-10-29 11:00:33.256668] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.792 [2024-10-29 11:00:33.259087] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.792 [2024-10-29 11:00:33.259128] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:27.792 BaseBdev2 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 BaseBdev3_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.792 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.792 true 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 [2024-10-29 11:00:33.298012] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:28.053 [2024-10-29 11:00:33.298061] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:28.053 [2024-10-29 11:00:33.298083] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:28.053 [2024-10-29 11:00:33.298094] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:28.053 [2024-10-29 11:00:33.300580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:28.053 [2024-10-29 11:00:33.300623] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:28.053 BaseBdev3 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 BaseBdev4_malloc 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 true 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 [2024-10-29 11:00:33.350603] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:28.053 [2024-10-29 11:00:33.350660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:28.053 [2024-10-29 11:00:33.350700] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:28.053 [2024-10-29 11:00:33.350710] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:28.053 [2024-10-29 11:00:33.353139] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:28.053 [2024-10-29 11:00:33.353181] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:28.053 BaseBdev4 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 [2024-10-29 11:00:33.362637] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:28.053 [2024-10-29 11:00:33.364840] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:28.053 [2024-10-29 11:00:33.364937] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:28.053 [2024-10-29 11:00:33.364998] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:28.053 [2024-10-29 11:00:33.365248] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:11:28.053 [2024-10-29 11:00:33.365271] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:28.053 [2024-10-29 11:00:33.365603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:11:28.053 [2024-10-29 11:00:33.365800] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:11:28.053 [2024-10-29 11:00:33.365834] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:11:28.053 [2024-10-29 11:00:33.365991] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.053 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.053 "name": "raid_bdev1", 00:11:28.053 "uuid": "e7a14551-4282-43a6-8c69-828118e6ede5", 00:11:28.053 "strip_size_kb": 0, 00:11:28.053 "state": "online", 00:11:28.053 "raid_level": "raid1", 00:11:28.053 "superblock": true, 00:11:28.053 "num_base_bdevs": 4, 00:11:28.053 "num_base_bdevs_discovered": 4, 00:11:28.053 "num_base_bdevs_operational": 4, 00:11:28.053 "base_bdevs_list": [ 00:11:28.053 { 00:11:28.053 "name": "BaseBdev1", 00:11:28.053 "uuid": "793ffbbd-a544-5ed4-8d06-365603785d6b", 00:11:28.053 "is_configured": true, 00:11:28.053 "data_offset": 2048, 00:11:28.053 "data_size": 63488 00:11:28.053 }, 00:11:28.053 { 00:11:28.053 "name": "BaseBdev2", 00:11:28.053 "uuid": "d595fbce-d130-5c94-8b59-aa3c3c9d3be4", 00:11:28.053 "is_configured": true, 00:11:28.053 "data_offset": 2048, 00:11:28.053 "data_size": 63488 00:11:28.053 }, 00:11:28.053 { 00:11:28.053 "name": "BaseBdev3", 00:11:28.053 "uuid": "a98a135c-22f0-592d-a679-f890bfa405fa", 00:11:28.053 "is_configured": true, 00:11:28.053 "data_offset": 2048, 00:11:28.053 "data_size": 63488 00:11:28.053 }, 00:11:28.053 { 00:11:28.053 "name": "BaseBdev4", 00:11:28.053 "uuid": "a49856f6-1926-5da7-b05e-af64a7d58b50", 00:11:28.053 "is_configured": true, 00:11:28.053 "data_offset": 2048, 00:11:28.053 "data_size": 63488 00:11:28.053 } 00:11:28.053 ] 00:11:28.053 }' 00:11:28.054 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.054 11:00:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.313 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:28.313 11:00:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:28.573 [2024-10-29 11:00:33.878195] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.511 "name": "raid_bdev1", 00:11:29.511 "uuid": "e7a14551-4282-43a6-8c69-828118e6ede5", 00:11:29.511 "strip_size_kb": 0, 00:11:29.511 "state": "online", 00:11:29.511 "raid_level": "raid1", 00:11:29.511 "superblock": true, 00:11:29.511 "num_base_bdevs": 4, 00:11:29.511 "num_base_bdevs_discovered": 4, 00:11:29.511 "num_base_bdevs_operational": 4, 00:11:29.511 "base_bdevs_list": [ 00:11:29.511 { 00:11:29.511 "name": "BaseBdev1", 00:11:29.511 "uuid": "793ffbbd-a544-5ed4-8d06-365603785d6b", 00:11:29.511 "is_configured": true, 00:11:29.511 "data_offset": 2048, 00:11:29.511 "data_size": 63488 00:11:29.511 }, 00:11:29.511 { 00:11:29.511 "name": "BaseBdev2", 00:11:29.511 "uuid": "d595fbce-d130-5c94-8b59-aa3c3c9d3be4", 00:11:29.511 "is_configured": true, 00:11:29.511 "data_offset": 2048, 00:11:29.511 "data_size": 63488 00:11:29.511 }, 00:11:29.511 { 00:11:29.511 "name": "BaseBdev3", 00:11:29.511 "uuid": "a98a135c-22f0-592d-a679-f890bfa405fa", 00:11:29.511 "is_configured": true, 00:11:29.511 "data_offset": 2048, 00:11:29.511 "data_size": 63488 00:11:29.511 }, 00:11:29.511 { 00:11:29.511 "name": "BaseBdev4", 00:11:29.511 "uuid": "a49856f6-1926-5da7-b05e-af64a7d58b50", 00:11:29.511 "is_configured": true, 00:11:29.511 "data_offset": 2048, 00:11:29.511 "data_size": 63488 00:11:29.511 } 00:11:29.511 ] 00:11:29.511 }' 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.511 11:00:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.771 [2024-10-29 11:00:35.236849] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:29.771 [2024-10-29 11:00:35.236886] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:29.771 [2024-10-29 11:00:35.239302] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:29.771 [2024-10-29 11:00:35.239370] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:29.771 [2024-10-29 11:00:35.239505] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:29.771 [2024-10-29 11:00:35.239615] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:11:29.771 { 00:11:29.771 "results": [ 00:11:29.771 { 00:11:29.771 "job": "raid_bdev1", 00:11:29.771 "core_mask": "0x1", 00:11:29.771 "workload": "randrw", 00:11:29.771 "percentage": 50, 00:11:29.771 "status": "finished", 00:11:29.771 "queue_depth": 1, 00:11:29.771 "io_size": 131072, 00:11:29.771 "runtime": 1.359001, 00:11:29.771 "iops": 10392.928334857737, 00:11:29.771 "mibps": 1299.1160418572172, 00:11:29.771 "io_failed": 0, 00:11:29.771 "io_timeout": 0, 00:11:29.771 "avg_latency_us": 93.23863324095132, 00:11:29.771 "min_latency_us": 23.02882096069869, 00:11:29.771 "max_latency_us": 1738.564192139738 00:11:29.771 } 00:11:29.771 ], 00:11:29.771 "core_count": 1 00:11:29.771 } 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85914 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@952 -- # '[' -z 85914 ']' 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # kill -0 85914 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # uname 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:29.771 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 85914 00:11:30.030 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:30.030 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:30.030 killing process with pid 85914 00:11:30.030 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 85914' 00:11:30.030 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@971 -- # kill 85914 00:11:30.030 [2024-10-29 11:00:35.295021] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:30.030 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@976 -- # wait 85914 00:11:30.030 [2024-10-29 11:00:35.332205] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.X8GMJiOjKE 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:30.289 00:11:30.289 real 0m3.354s 00:11:30.289 user 0m4.252s 00:11:30.289 sys 0m0.545s 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:30.289 11:00:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:30.289 ************************************ 00:11:30.289 END TEST raid_read_error_test 00:11:30.289 ************************************ 00:11:30.289 11:00:35 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:11:30.289 11:00:35 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:11:30.289 11:00:35 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:30.289 11:00:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:30.289 ************************************ 00:11:30.289 START TEST raid_write_error_test 00:11:30.289 ************************************ 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1127 -- # raid_io_error_test raid1 4 write 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.z2GnDUMM1h 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=86043 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 86043 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@833 -- # '[' -z 86043 ']' 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:30.289 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:30.289 11:00:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:30.289 [2024-10-29 11:00:35.734961] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:30.289 [2024-10-29 11:00:35.735104] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86043 ] 00:11:30.549 [2024-10-29 11:00:35.892397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:30.549 [2024-10-29 11:00:35.917552] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:30.549 [2024-10-29 11:00:35.960501] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:30.549 [2024-10-29 11:00:35.960538] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@866 -- # return 0 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 BaseBdev1_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 true 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 [2024-10-29 11:00:36.575189] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:31.140 [2024-10-29 11:00:36.575240] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.140 [2024-10-29 11:00:36.575262] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:11:31.140 [2024-10-29 11:00:36.575290] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.140 [2024-10-29 11:00:36.577448] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.140 [2024-10-29 11:00:36.577485] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:31.140 BaseBdev1 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 BaseBdev2_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 true 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 [2024-10-29 11:00:36.603839] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:31.140 [2024-10-29 11:00:36.603890] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.140 [2024-10-29 11:00:36.603909] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:31.140 [2024-10-29 11:00:36.603917] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.140 [2024-10-29 11:00:36.605898] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.140 [2024-10-29 11:00:36.605934] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:31.140 BaseBdev2 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.140 BaseBdev3_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.140 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 true 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 [2024-10-29 11:00:36.644641] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:31.405 [2024-10-29 11:00:36.644693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.405 [2024-10-29 11:00:36.644713] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:11:31.405 [2024-10-29 11:00:36.644722] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.405 [2024-10-29 11:00:36.646800] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.405 [2024-10-29 11:00:36.646836] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:31.405 BaseBdev3 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 BaseBdev4_malloc 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 true 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 [2024-10-29 11:00:36.694816] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:31.405 [2024-10-29 11:00:36.694865] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.405 [2024-10-29 11:00:36.694889] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:31.405 [2024-10-29 11:00:36.694898] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.405 [2024-10-29 11:00:36.697020] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.405 [2024-10-29 11:00:36.697059] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:31.405 BaseBdev4 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 [2024-10-29 11:00:36.706862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:31.405 [2024-10-29 11:00:36.708988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:31.405 [2024-10-29 11:00:36.709087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:31.405 [2024-10-29 11:00:36.709147] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:31.405 [2024-10-29 11:00:36.709428] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007080 00:11:31.405 [2024-10-29 11:00:36.709452] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:31.405 [2024-10-29 11:00:36.709746] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:11:31.405 [2024-10-29 11:00:36.709917] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007080 00:11:31.405 [2024-10-29 11:00:36.709955] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007080 00:11:31.405 [2024-10-29 11:00:36.710103] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.405 "name": "raid_bdev1", 00:11:31.405 "uuid": "e72a6fef-d326-44e0-a1cd-a153bc10253e", 00:11:31.405 "strip_size_kb": 0, 00:11:31.405 "state": "online", 00:11:31.405 "raid_level": "raid1", 00:11:31.405 "superblock": true, 00:11:31.405 "num_base_bdevs": 4, 00:11:31.405 "num_base_bdevs_discovered": 4, 00:11:31.405 "num_base_bdevs_operational": 4, 00:11:31.405 "base_bdevs_list": [ 00:11:31.405 { 00:11:31.405 "name": "BaseBdev1", 00:11:31.405 "uuid": "9177042c-c351-5afc-b3a5-8dc168fcfc0f", 00:11:31.405 "is_configured": true, 00:11:31.405 "data_offset": 2048, 00:11:31.405 "data_size": 63488 00:11:31.405 }, 00:11:31.405 { 00:11:31.405 "name": "BaseBdev2", 00:11:31.405 "uuid": "b3ac4460-804c-56a7-b5ab-3d50369f656b", 00:11:31.405 "is_configured": true, 00:11:31.405 "data_offset": 2048, 00:11:31.405 "data_size": 63488 00:11:31.405 }, 00:11:31.405 { 00:11:31.405 "name": "BaseBdev3", 00:11:31.405 "uuid": "3856eb91-32a1-5519-976f-bafff25e5d00", 00:11:31.405 "is_configured": true, 00:11:31.405 "data_offset": 2048, 00:11:31.405 "data_size": 63488 00:11:31.405 }, 00:11:31.405 { 00:11:31.405 "name": "BaseBdev4", 00:11:31.405 "uuid": "db8334e5-1b27-535e-a382-7ab4cae5eac8", 00:11:31.405 "is_configured": true, 00:11:31.405 "data_offset": 2048, 00:11:31.405 "data_size": 63488 00:11:31.405 } 00:11:31.405 ] 00:11:31.405 }' 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.405 11:00:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.664 11:00:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:31.664 11:00:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:31.923 [2024-10-29 11:00:37.226303] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.863 [2024-10-29 11:00:38.145093] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:11:32.863 [2024-10-29 11:00:38.145223] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:32.863 [2024-10-29 11:00:38.145494] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.863 "name": "raid_bdev1", 00:11:32.863 "uuid": "e72a6fef-d326-44e0-a1cd-a153bc10253e", 00:11:32.863 "strip_size_kb": 0, 00:11:32.863 "state": "online", 00:11:32.863 "raid_level": "raid1", 00:11:32.863 "superblock": true, 00:11:32.863 "num_base_bdevs": 4, 00:11:32.863 "num_base_bdevs_discovered": 3, 00:11:32.863 "num_base_bdevs_operational": 3, 00:11:32.863 "base_bdevs_list": [ 00:11:32.863 { 00:11:32.863 "name": null, 00:11:32.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:32.863 "is_configured": false, 00:11:32.863 "data_offset": 0, 00:11:32.863 "data_size": 63488 00:11:32.863 }, 00:11:32.863 { 00:11:32.863 "name": "BaseBdev2", 00:11:32.863 "uuid": "b3ac4460-804c-56a7-b5ab-3d50369f656b", 00:11:32.863 "is_configured": true, 00:11:32.863 "data_offset": 2048, 00:11:32.863 "data_size": 63488 00:11:32.863 }, 00:11:32.863 { 00:11:32.863 "name": "BaseBdev3", 00:11:32.863 "uuid": "3856eb91-32a1-5519-976f-bafff25e5d00", 00:11:32.863 "is_configured": true, 00:11:32.863 "data_offset": 2048, 00:11:32.863 "data_size": 63488 00:11:32.863 }, 00:11:32.863 { 00:11:32.863 "name": "BaseBdev4", 00:11:32.863 "uuid": "db8334e5-1b27-535e-a382-7ab4cae5eac8", 00:11:32.863 "is_configured": true, 00:11:32.863 "data_offset": 2048, 00:11:32.863 "data_size": 63488 00:11:32.863 } 00:11:32.863 ] 00:11:32.863 }' 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.863 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.125 [2024-10-29 11:00:38.536163] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:33.125 [2024-10-29 11:00:38.536199] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:33.125 [2024-10-29 11:00:38.538934] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:33.125 [2024-10-29 11:00:38.538983] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:33.125 [2024-10-29 11:00:38.539080] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:33.125 [2024-10-29 11:00:38.539091] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state offline 00:11:33.125 { 00:11:33.125 "results": [ 00:11:33.125 { 00:11:33.125 "job": "raid_bdev1", 00:11:33.125 "core_mask": "0x1", 00:11:33.125 "workload": "randrw", 00:11:33.125 "percentage": 50, 00:11:33.125 "status": "finished", 00:11:33.125 "queue_depth": 1, 00:11:33.125 "io_size": 131072, 00:11:33.125 "runtime": 1.31051, 00:11:33.125 "iops": 12444.010347116771, 00:11:33.125 "mibps": 1555.5012933895964, 00:11:33.125 "io_failed": 0, 00:11:33.125 "io_timeout": 0, 00:11:33.125 "avg_latency_us": 77.81859638637451, 00:11:33.125 "min_latency_us": 22.581659388646287, 00:11:33.125 "max_latency_us": 1423.7624454148472 00:11:33.125 } 00:11:33.125 ], 00:11:33.125 "core_count": 1 00:11:33.125 } 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 86043 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@952 -- # '[' -z 86043 ']' 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # kill -0 86043 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # uname 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 86043 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:33.125 killing process with pid 86043 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 86043' 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@971 -- # kill 86043 00:11:33.125 [2024-10-29 11:00:38.586124] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:33.125 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@976 -- # wait 86043 00:11:33.385 [2024-10-29 11:00:38.621102] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.z2GnDUMM1h 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:33.385 ************************************ 00:11:33.385 END TEST raid_write_error_test 00:11:33.385 ************************************ 00:11:33.385 00:11:33.385 real 0m3.221s 00:11:33.385 user 0m4.009s 00:11:33.385 sys 0m0.543s 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:33.385 11:00:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.645 11:00:38 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:11:33.645 11:00:38 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:33.645 11:00:38 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:11:33.645 11:00:38 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:11:33.645 11:00:38 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:33.645 11:00:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:33.645 ************************************ 00:11:33.645 START TEST raid_rebuild_test 00:11:33.645 ************************************ 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 false false true 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=86176 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 86176 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@833 -- # '[' -z 86176 ']' 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:33.645 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:33.645 11:00:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.645 [2024-10-29 11:00:39.021363] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:33.645 [2024-10-29 11:00:39.021597] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:11:33.645 Zero copy mechanism will not be used. 00:11:33.645 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86176 ] 00:11:33.905 [2024-10-29 11:00:39.199415] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:33.905 [2024-10-29 11:00:39.226889] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:33.905 [2024-10-29 11:00:39.271532] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:33.905 [2024-10-29 11:00:39.271650] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # return 0 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 BaseBdev1_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 [2024-10-29 11:00:39.851093] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:34.476 [2024-10-29 11:00:39.851200] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.476 [2024-10-29 11:00:39.851246] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:34.476 [2024-10-29 11:00:39.851280] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.476 [2024-10-29 11:00:39.853523] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.476 [2024-10-29 11:00:39.853593] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:34.476 BaseBdev1 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 BaseBdev2_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 [2024-10-29 11:00:39.871784] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:34.476 [2024-10-29 11:00:39.871838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.476 [2024-10-29 11:00:39.871858] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:34.476 [2024-10-29 11:00:39.871867] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.476 [2024-10-29 11:00:39.874189] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.476 [2024-10-29 11:00:39.874225] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:34.476 BaseBdev2 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 spare_malloc 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 spare_delay 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 [2024-10-29 11:00:39.900783] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:34.476 [2024-10-29 11:00:39.900843] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.476 [2024-10-29 11:00:39.900866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:34.476 [2024-10-29 11:00:39.900875] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.476 [2024-10-29 11:00:39.903038] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.476 [2024-10-29 11:00:39.903075] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:34.476 spare 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 [2024-10-29 11:00:39.908821] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:34.476 [2024-10-29 11:00:39.911003] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:34.476 [2024-10-29 11:00:39.911192] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:34.476 [2024-10-29 11:00:39.911219] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:34.476 [2024-10-29 11:00:39.911566] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:11:34.476 [2024-10-29 11:00:39.911724] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:34.476 [2024-10-29 11:00:39.911742] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:34.476 [2024-10-29 11:00:39.911872] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.476 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.476 "name": "raid_bdev1", 00:11:34.476 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:34.476 "strip_size_kb": 0, 00:11:34.476 "state": "online", 00:11:34.476 "raid_level": "raid1", 00:11:34.476 "superblock": false, 00:11:34.476 "num_base_bdevs": 2, 00:11:34.476 "num_base_bdevs_discovered": 2, 00:11:34.476 "num_base_bdevs_operational": 2, 00:11:34.476 "base_bdevs_list": [ 00:11:34.476 { 00:11:34.476 "name": "BaseBdev1", 00:11:34.476 "uuid": "082a9643-4b7d-5afc-8f67-87875fdb005b", 00:11:34.476 "is_configured": true, 00:11:34.476 "data_offset": 0, 00:11:34.476 "data_size": 65536 00:11:34.476 }, 00:11:34.476 { 00:11:34.476 "name": "BaseBdev2", 00:11:34.476 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:34.476 "is_configured": true, 00:11:34.476 "data_offset": 0, 00:11:34.477 "data_size": 65536 00:11:34.477 } 00:11:34.477 ] 00:11:34.477 }' 00:11:34.477 11:00:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.477 11:00:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.046 [2024-10-29 11:00:40.368286] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:35.046 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:35.306 [2024-10-29 11:00:40.631638] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:35.306 /dev/nbd0 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:35.306 1+0 records in 00:11:35.306 1+0 records out 00:11:35.306 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000554104 s, 7.4 MB/s 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:35.306 11:00:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:39.505 65536+0 records in 00:11:39.505 65536+0 records out 00:11:39.505 33554432 bytes (34 MB, 32 MiB) copied, 3.55554 s, 9.4 MB/s 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:39.505 [2024-10-29 11:00:44.481113] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:39.505 11:00:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.506 [2024-10-29 11:00:44.497172] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.506 "name": "raid_bdev1", 00:11:39.506 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:39.506 "strip_size_kb": 0, 00:11:39.506 "state": "online", 00:11:39.506 "raid_level": "raid1", 00:11:39.506 "superblock": false, 00:11:39.506 "num_base_bdevs": 2, 00:11:39.506 "num_base_bdevs_discovered": 1, 00:11:39.506 "num_base_bdevs_operational": 1, 00:11:39.506 "base_bdevs_list": [ 00:11:39.506 { 00:11:39.506 "name": null, 00:11:39.506 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:39.506 "is_configured": false, 00:11:39.506 "data_offset": 0, 00:11:39.506 "data_size": 65536 00:11:39.506 }, 00:11:39.506 { 00:11:39.506 "name": "BaseBdev2", 00:11:39.506 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:39.506 "is_configured": true, 00:11:39.506 "data_offset": 0, 00:11:39.506 "data_size": 65536 00:11:39.506 } 00:11:39.506 ] 00:11:39.506 }' 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.506 [2024-10-29 11:00:44.968439] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:39.506 [2024-10-29 11:00:44.984384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09a30 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.506 11:00:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:39.506 [2024-10-29 11:00:44.986767] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:40.888 11:00:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.888 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:40.888 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:40.888 "name": "raid_bdev1", 00:11:40.888 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:40.888 "strip_size_kb": 0, 00:11:40.888 "state": "online", 00:11:40.888 "raid_level": "raid1", 00:11:40.888 "superblock": false, 00:11:40.888 "num_base_bdevs": 2, 00:11:40.888 "num_base_bdevs_discovered": 2, 00:11:40.888 "num_base_bdevs_operational": 2, 00:11:40.888 "process": { 00:11:40.888 "type": "rebuild", 00:11:40.888 "target": "spare", 00:11:40.888 "progress": { 00:11:40.888 "blocks": 20480, 00:11:40.888 "percent": 31 00:11:40.888 } 00:11:40.888 }, 00:11:40.888 "base_bdevs_list": [ 00:11:40.888 { 00:11:40.888 "name": "spare", 00:11:40.888 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:40.889 "is_configured": true, 00:11:40.889 "data_offset": 0, 00:11:40.889 "data_size": 65536 00:11:40.889 }, 00:11:40.889 { 00:11:40.889 "name": "BaseBdev2", 00:11:40.889 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:40.889 "is_configured": true, 00:11:40.889 "data_offset": 0, 00:11:40.889 "data_size": 65536 00:11:40.889 } 00:11:40.889 ] 00:11:40.889 }' 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.889 [2024-10-29 11:00:46.127708] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:40.889 [2024-10-29 11:00:46.191646] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:40.889 [2024-10-29 11:00:46.191745] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:40.889 [2024-10-29 11:00:46.191766] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:40.889 [2024-10-29 11:00:46.191774] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:40.889 "name": "raid_bdev1", 00:11:40.889 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:40.889 "strip_size_kb": 0, 00:11:40.889 "state": "online", 00:11:40.889 "raid_level": "raid1", 00:11:40.889 "superblock": false, 00:11:40.889 "num_base_bdevs": 2, 00:11:40.889 "num_base_bdevs_discovered": 1, 00:11:40.889 "num_base_bdevs_operational": 1, 00:11:40.889 "base_bdevs_list": [ 00:11:40.889 { 00:11:40.889 "name": null, 00:11:40.889 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:40.889 "is_configured": false, 00:11:40.889 "data_offset": 0, 00:11:40.889 "data_size": 65536 00:11:40.889 }, 00:11:40.889 { 00:11:40.889 "name": "BaseBdev2", 00:11:40.889 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:40.889 "is_configured": true, 00:11:40.889 "data_offset": 0, 00:11:40.889 "data_size": 65536 00:11:40.889 } 00:11:40.889 ] 00:11:40.889 }' 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:40.889 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.149 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.410 "name": "raid_bdev1", 00:11:41.410 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:41.410 "strip_size_kb": 0, 00:11:41.410 "state": "online", 00:11:41.410 "raid_level": "raid1", 00:11:41.410 "superblock": false, 00:11:41.410 "num_base_bdevs": 2, 00:11:41.410 "num_base_bdevs_discovered": 1, 00:11:41.410 "num_base_bdevs_operational": 1, 00:11:41.410 "base_bdevs_list": [ 00:11:41.410 { 00:11:41.410 "name": null, 00:11:41.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:41.410 "is_configured": false, 00:11:41.410 "data_offset": 0, 00:11:41.410 "data_size": 65536 00:11:41.410 }, 00:11:41.410 { 00:11:41.410 "name": "BaseBdev2", 00:11:41.410 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:41.410 "is_configured": true, 00:11:41.410 "data_offset": 0, 00:11:41.410 "data_size": 65536 00:11:41.410 } 00:11:41.410 ] 00:11:41.410 }' 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.410 [2024-10-29 11:00:46.759901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:41.410 [2024-10-29 11:00:46.764995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09b00 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.410 11:00:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:41.410 [2024-10-29 11:00:46.766954] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:42.351 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.352 "name": "raid_bdev1", 00:11:42.352 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:42.352 "strip_size_kb": 0, 00:11:42.352 "state": "online", 00:11:42.352 "raid_level": "raid1", 00:11:42.352 "superblock": false, 00:11:42.352 "num_base_bdevs": 2, 00:11:42.352 "num_base_bdevs_discovered": 2, 00:11:42.352 "num_base_bdevs_operational": 2, 00:11:42.352 "process": { 00:11:42.352 "type": "rebuild", 00:11:42.352 "target": "spare", 00:11:42.352 "progress": { 00:11:42.352 "blocks": 20480, 00:11:42.352 "percent": 31 00:11:42.352 } 00:11:42.352 }, 00:11:42.352 "base_bdevs_list": [ 00:11:42.352 { 00:11:42.352 "name": "spare", 00:11:42.352 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:42.352 "is_configured": true, 00:11:42.352 "data_offset": 0, 00:11:42.352 "data_size": 65536 00:11:42.352 }, 00:11:42.352 { 00:11:42.352 "name": "BaseBdev2", 00:11:42.352 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:42.352 "is_configured": true, 00:11:42.352 "data_offset": 0, 00:11:42.352 "data_size": 65536 00:11:42.352 } 00:11:42.352 ] 00:11:42.352 }' 00:11:42.352 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=292 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.612 "name": "raid_bdev1", 00:11:42.612 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:42.612 "strip_size_kb": 0, 00:11:42.612 "state": "online", 00:11:42.612 "raid_level": "raid1", 00:11:42.612 "superblock": false, 00:11:42.612 "num_base_bdevs": 2, 00:11:42.612 "num_base_bdevs_discovered": 2, 00:11:42.612 "num_base_bdevs_operational": 2, 00:11:42.612 "process": { 00:11:42.612 "type": "rebuild", 00:11:42.612 "target": "spare", 00:11:42.612 "progress": { 00:11:42.612 "blocks": 22528, 00:11:42.612 "percent": 34 00:11:42.612 } 00:11:42.612 }, 00:11:42.612 "base_bdevs_list": [ 00:11:42.612 { 00:11:42.612 "name": "spare", 00:11:42.612 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:42.612 "is_configured": true, 00:11:42.612 "data_offset": 0, 00:11:42.612 "data_size": 65536 00:11:42.612 }, 00:11:42.612 { 00:11:42.612 "name": "BaseBdev2", 00:11:42.612 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:42.612 "is_configured": true, 00:11:42.612 "data_offset": 0, 00:11:42.612 "data_size": 65536 00:11:42.612 } 00:11:42.612 ] 00:11:42.612 }' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.612 11:00:47 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.612 11:00:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.612 11:00:48 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.621 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.622 11:00:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:43.622 11:00:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.622 11:00:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:43.622 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:43.622 "name": "raid_bdev1", 00:11:43.622 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:43.622 "strip_size_kb": 0, 00:11:43.622 "state": "online", 00:11:43.622 "raid_level": "raid1", 00:11:43.622 "superblock": false, 00:11:43.622 "num_base_bdevs": 2, 00:11:43.622 "num_base_bdevs_discovered": 2, 00:11:43.622 "num_base_bdevs_operational": 2, 00:11:43.622 "process": { 00:11:43.622 "type": "rebuild", 00:11:43.622 "target": "spare", 00:11:43.622 "progress": { 00:11:43.622 "blocks": 45056, 00:11:43.622 "percent": 68 00:11:43.622 } 00:11:43.622 }, 00:11:43.622 "base_bdevs_list": [ 00:11:43.622 { 00:11:43.622 "name": "spare", 00:11:43.622 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:43.622 "is_configured": true, 00:11:43.622 "data_offset": 0, 00:11:43.622 "data_size": 65536 00:11:43.622 }, 00:11:43.622 { 00:11:43.622 "name": "BaseBdev2", 00:11:43.622 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:43.622 "is_configured": true, 00:11:43.622 "data_offset": 0, 00:11:43.622 "data_size": 65536 00:11:43.622 } 00:11:43.622 ] 00:11:43.622 }' 00:11:43.622 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:43.882 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:43.882 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:43.882 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:43.882 11:00:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:44.820 [2024-10-29 11:00:49.978116] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:44.820 [2024-10-29 11:00:49.978258] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:44.820 [2024-10-29 11:00:49.978351] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.820 "name": "raid_bdev1", 00:11:44.820 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:44.820 "strip_size_kb": 0, 00:11:44.820 "state": "online", 00:11:44.820 "raid_level": "raid1", 00:11:44.820 "superblock": false, 00:11:44.820 "num_base_bdevs": 2, 00:11:44.820 "num_base_bdevs_discovered": 2, 00:11:44.820 "num_base_bdevs_operational": 2, 00:11:44.820 "base_bdevs_list": [ 00:11:44.820 { 00:11:44.820 "name": "spare", 00:11:44.820 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:44.820 "is_configured": true, 00:11:44.820 "data_offset": 0, 00:11:44.820 "data_size": 65536 00:11:44.820 }, 00:11:44.820 { 00:11:44.820 "name": "BaseBdev2", 00:11:44.820 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:44.820 "is_configured": true, 00:11:44.820 "data_offset": 0, 00:11:44.820 "data_size": 65536 00:11:44.820 } 00:11:44.820 ] 00:11:44.820 }' 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:44.820 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.080 "name": "raid_bdev1", 00:11:45.080 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:45.080 "strip_size_kb": 0, 00:11:45.080 "state": "online", 00:11:45.080 "raid_level": "raid1", 00:11:45.080 "superblock": false, 00:11:45.080 "num_base_bdevs": 2, 00:11:45.080 "num_base_bdevs_discovered": 2, 00:11:45.080 "num_base_bdevs_operational": 2, 00:11:45.080 "base_bdevs_list": [ 00:11:45.080 { 00:11:45.080 "name": "spare", 00:11:45.080 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 }, 00:11:45.080 { 00:11:45.080 "name": "BaseBdev2", 00:11:45.080 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 } 00:11:45.080 ] 00:11:45.080 }' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.080 "name": "raid_bdev1", 00:11:45.080 "uuid": "176d4a24-cb8a-48a0-b5dc-2966a2d9a775", 00:11:45.080 "strip_size_kb": 0, 00:11:45.080 "state": "online", 00:11:45.080 "raid_level": "raid1", 00:11:45.080 "superblock": false, 00:11:45.080 "num_base_bdevs": 2, 00:11:45.080 "num_base_bdevs_discovered": 2, 00:11:45.080 "num_base_bdevs_operational": 2, 00:11:45.080 "base_bdevs_list": [ 00:11:45.080 { 00:11:45.080 "name": "spare", 00:11:45.080 "uuid": "3dd88e5a-57e0-5ce5-be13-545b74dec8fd", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 }, 00:11:45.080 { 00:11:45.080 "name": "BaseBdev2", 00:11:45.080 "uuid": "8d742dd0-8e86-5675-b4cc-eaa02147e8de", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 } 00:11:45.080 ] 00:11:45.080 }' 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.080 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.651 [2024-10-29 11:00:50.861678] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:45.651 [2024-10-29 11:00:50.861756] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:45.651 [2024-10-29 11:00:50.861865] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:45.651 [2024-10-29 11:00:50.861982] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:45.651 [2024-10-29 11:00:50.862051] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:45.651 11:00:50 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:45.651 /dev/nbd0 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:45.651 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:45.912 1+0 records in 00:11:45.912 1+0 records out 00:11:45.912 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000500323 s, 8.2 MB/s 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:45.912 /dev/nbd1 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:45.912 1+0 records in 00:11:45.912 1+0 records out 00:11:45.912 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000338705 s, 12.1 MB/s 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:45.912 11:00:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:46.172 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 86176 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@952 -- # '[' -z 86176 ']' 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # kill -0 86176 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@957 -- # uname 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:11:46.432 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 86176 00:11:46.692 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:11:46.692 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:11:46.692 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 86176' 00:11:46.692 killing process with pid 86176 00:11:46.692 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@971 -- # kill 86176 00:11:46.692 Received shutdown signal, test time was about 60.000000 seconds 00:11:46.692 00:11:46.692 Latency(us) 00:11:46.692 [2024-10-29T11:00:52.190Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:46.692 [2024-10-29T11:00:52.190Z] =================================================================================================================== 00:11:46.692 [2024-10-29T11:00:52.190Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:46.692 [2024-10-29 11:00:51.945042] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:46.692 11:00:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@976 -- # wait 86176 00:11:46.692 [2024-10-29 11:00:51.976470] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:46.692 11:00:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:46.692 00:11:46.692 real 0m13.267s 00:11:46.692 user 0m15.486s 00:11:46.692 sys 0m2.713s 00:11:46.692 ************************************ 00:11:46.692 END TEST raid_rebuild_test 00:11:46.692 ************************************ 00:11:46.692 11:00:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:11:46.692 11:00:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:46.953 11:00:52 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:46.953 11:00:52 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:11:46.953 11:00:52 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:11:46.953 11:00:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:46.953 ************************************ 00:11:46.953 START TEST raid_rebuild_test_sb 00:11:46.953 ************************************ 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 true false true 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86572 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86572 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@833 -- # '[' -z 86572 ']' 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:46.953 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:11:46.953 11:00:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:46.953 [2024-10-29 11:00:52.351037] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:11:46.953 [2024-10-29 11:00:52.351261] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:11:46.953 Zero copy mechanism will not be used. 00:11:46.953 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86572 ] 00:11:47.212 [2024-10-29 11:00:52.523017] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:47.212 [2024-10-29 11:00:52.548397] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:47.212 [2024-10-29 11:00:52.591934] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:47.212 [2024-10-29 11:00:52.592050] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # return 0 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 BaseBdev1_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 [2024-10-29 11:00:53.187542] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:47.783 [2024-10-29 11:00:53.187643] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:47.783 [2024-10-29 11:00:53.187699] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:47.783 [2024-10-29 11:00:53.187720] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:47.783 [2024-10-29 11:00:53.189905] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:47.783 [2024-10-29 11:00:53.189946] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:47.783 BaseBdev1 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 BaseBdev2_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 [2024-10-29 11:00:53.216303] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:47.783 [2024-10-29 11:00:53.216383] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:47.783 [2024-10-29 11:00:53.216406] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:47.783 [2024-10-29 11:00:53.216414] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:47.783 [2024-10-29 11:00:53.218440] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:47.783 [2024-10-29 11:00:53.218510] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:47.783 BaseBdev2 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 spare_malloc 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 spare_delay 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 [2024-10-29 11:00:53.256997] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:47.783 [2024-10-29 11:00:53.257048] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:47.783 [2024-10-29 11:00:53.257069] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:11:47.783 [2024-10-29 11:00:53.257078] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:47.783 [2024-10-29 11:00:53.259190] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:47.783 [2024-10-29 11:00:53.259226] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:47.783 spare 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.783 [2024-10-29 11:00:53.269006] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:47.783 [2024-10-29 11:00:53.270849] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:47.783 [2024-10-29 11:00:53.270999] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:11:47.783 [2024-10-29 11:00:53.271011] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:47.783 [2024-10-29 11:00:53.271240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:11:47.783 [2024-10-29 11:00:53.271377] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:11:47.783 [2024-10-29 11:00:53.271408] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:11:47.783 [2024-10-29 11:00:53.271516] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:47.783 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.044 "name": "raid_bdev1", 00:11:48.044 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:48.044 "strip_size_kb": 0, 00:11:48.044 "state": "online", 00:11:48.044 "raid_level": "raid1", 00:11:48.044 "superblock": true, 00:11:48.044 "num_base_bdevs": 2, 00:11:48.044 "num_base_bdevs_discovered": 2, 00:11:48.044 "num_base_bdevs_operational": 2, 00:11:48.044 "base_bdevs_list": [ 00:11:48.044 { 00:11:48.044 "name": "BaseBdev1", 00:11:48.044 "uuid": "4e758927-609a-596b-9881-46050c1ecc40", 00:11:48.044 "is_configured": true, 00:11:48.044 "data_offset": 2048, 00:11:48.044 "data_size": 63488 00:11:48.044 }, 00:11:48.044 { 00:11:48.044 "name": "BaseBdev2", 00:11:48.044 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:48.044 "is_configured": true, 00:11:48.044 "data_offset": 2048, 00:11:48.044 "data_size": 63488 00:11:48.044 } 00:11:48.044 ] 00:11:48.044 }' 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.044 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.303 [2024-10-29 11:00:53.704542] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:48.303 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:48.563 [2024-10-29 11:00:53.943929] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:11:48.563 /dev/nbd0 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:48.563 1+0 records in 00:11:48.563 1+0 records out 00:11:48.563 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000488311 s, 8.4 MB/s 00:11:48.563 11:00:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:48.563 11:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:11:48.563 11:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:48.563 11:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:48.563 11:00:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:11:48.564 11:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:48.564 11:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:48.564 11:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:48.564 11:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:48.564 11:00:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:52.754 63488+0 records in 00:11:52.754 63488+0 records out 00:11:52.754 32505856 bytes (33 MB, 31 MiB) copied, 3.60885 s, 9.0 MB/s 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:52.754 [2024-10-29 11:00:57.832462] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.754 [2024-10-29 11:00:57.852102] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.754 "name": "raid_bdev1", 00:11:52.754 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:52.754 "strip_size_kb": 0, 00:11:52.754 "state": "online", 00:11:52.754 "raid_level": "raid1", 00:11:52.754 "superblock": true, 00:11:52.754 "num_base_bdevs": 2, 00:11:52.754 "num_base_bdevs_discovered": 1, 00:11:52.754 "num_base_bdevs_operational": 1, 00:11:52.754 "base_bdevs_list": [ 00:11:52.754 { 00:11:52.754 "name": null, 00:11:52.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.754 "is_configured": false, 00:11:52.754 "data_offset": 0, 00:11:52.754 "data_size": 63488 00:11:52.754 }, 00:11:52.754 { 00:11:52.754 "name": "BaseBdev2", 00:11:52.754 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:52.754 "is_configured": true, 00:11:52.754 "data_offset": 2048, 00:11:52.754 "data_size": 63488 00:11:52.754 } 00:11:52.754 ] 00:11:52.754 }' 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.754 11:00:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.015 11:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:53.015 11:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.015 11:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.015 [2024-10-29 11:00:58.323445] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:53.015 [2024-10-29 11:00:58.339300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca31c0 00:11:53.015 11:00:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.015 11:00:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:53.015 [2024-10-29 11:00:58.341892] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:53.956 "name": "raid_bdev1", 00:11:53.956 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:53.956 "strip_size_kb": 0, 00:11:53.956 "state": "online", 00:11:53.956 "raid_level": "raid1", 00:11:53.956 "superblock": true, 00:11:53.956 "num_base_bdevs": 2, 00:11:53.956 "num_base_bdevs_discovered": 2, 00:11:53.956 "num_base_bdevs_operational": 2, 00:11:53.956 "process": { 00:11:53.956 "type": "rebuild", 00:11:53.956 "target": "spare", 00:11:53.956 "progress": { 00:11:53.956 "blocks": 20480, 00:11:53.956 "percent": 32 00:11:53.956 } 00:11:53.956 }, 00:11:53.956 "base_bdevs_list": [ 00:11:53.956 { 00:11:53.956 "name": "spare", 00:11:53.956 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:53.956 "is_configured": true, 00:11:53.956 "data_offset": 2048, 00:11:53.956 "data_size": 63488 00:11:53.956 }, 00:11:53.956 { 00:11:53.956 "name": "BaseBdev2", 00:11:53.956 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:53.956 "is_configured": true, 00:11:53.956 "data_offset": 2048, 00:11:53.956 "data_size": 63488 00:11:53.956 } 00:11:53.956 ] 00:11:53.956 }' 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:53.956 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:54.223 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:54.223 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:54.223 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.223 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.224 [2024-10-29 11:00:59.501073] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:54.224 [2024-10-29 11:00:59.546524] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:54.224 [2024-10-29 11:00:59.546616] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:54.224 [2024-10-29 11:00:59.546670] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:54.224 [2024-10-29 11:00:59.546691] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:54.224 "name": "raid_bdev1", 00:11:54.224 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:54.224 "strip_size_kb": 0, 00:11:54.224 "state": "online", 00:11:54.224 "raid_level": "raid1", 00:11:54.224 "superblock": true, 00:11:54.224 "num_base_bdevs": 2, 00:11:54.224 "num_base_bdevs_discovered": 1, 00:11:54.224 "num_base_bdevs_operational": 1, 00:11:54.224 "base_bdevs_list": [ 00:11:54.224 { 00:11:54.224 "name": null, 00:11:54.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.224 "is_configured": false, 00:11:54.224 "data_offset": 0, 00:11:54.224 "data_size": 63488 00:11:54.224 }, 00:11:54.224 { 00:11:54.224 "name": "BaseBdev2", 00:11:54.224 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:54.224 "is_configured": true, 00:11:54.224 "data_offset": 2048, 00:11:54.224 "data_size": 63488 00:11:54.224 } 00:11:54.224 ] 00:11:54.224 }' 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:54.224 11:00:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.796 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:54.796 11:00:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:54.796 "name": "raid_bdev1", 00:11:54.796 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:54.796 "strip_size_kb": 0, 00:11:54.796 "state": "online", 00:11:54.796 "raid_level": "raid1", 00:11:54.796 "superblock": true, 00:11:54.796 "num_base_bdevs": 2, 00:11:54.796 "num_base_bdevs_discovered": 1, 00:11:54.796 "num_base_bdevs_operational": 1, 00:11:54.796 "base_bdevs_list": [ 00:11:54.796 { 00:11:54.796 "name": null, 00:11:54.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.796 "is_configured": false, 00:11:54.796 "data_offset": 0, 00:11:54.796 "data_size": 63488 00:11:54.796 }, 00:11:54.796 { 00:11:54.796 "name": "BaseBdev2", 00:11:54.796 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:54.796 "is_configured": true, 00:11:54.796 "data_offset": 2048, 00:11:54.796 "data_size": 63488 00:11:54.796 } 00:11:54.796 ] 00:11:54.796 }' 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:54.796 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:54.797 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:54.797 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.797 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:54.797 [2024-10-29 11:01:00.126916] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:54.797 [2024-10-29 11:01:00.131692] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3290 00:11:54.797 11:01:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:54.797 11:01:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:54.797 [2024-10-29 11:01:00.133522] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.736 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.736 "name": "raid_bdev1", 00:11:55.736 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:55.736 "strip_size_kb": 0, 00:11:55.736 "state": "online", 00:11:55.736 "raid_level": "raid1", 00:11:55.736 "superblock": true, 00:11:55.736 "num_base_bdevs": 2, 00:11:55.736 "num_base_bdevs_discovered": 2, 00:11:55.736 "num_base_bdevs_operational": 2, 00:11:55.736 "process": { 00:11:55.736 "type": "rebuild", 00:11:55.736 "target": "spare", 00:11:55.736 "progress": { 00:11:55.736 "blocks": 20480, 00:11:55.736 "percent": 32 00:11:55.736 } 00:11:55.736 }, 00:11:55.737 "base_bdevs_list": [ 00:11:55.737 { 00:11:55.737 "name": "spare", 00:11:55.737 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:55.737 "is_configured": true, 00:11:55.737 "data_offset": 2048, 00:11:55.737 "data_size": 63488 00:11:55.737 }, 00:11:55.737 { 00:11:55.737 "name": "BaseBdev2", 00:11:55.737 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:55.737 "is_configured": true, 00:11:55.737 "data_offset": 2048, 00:11:55.737 "data_size": 63488 00:11:55.737 } 00:11:55.737 ] 00:11:55.737 }' 00:11:55.737 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:55.998 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=306 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.998 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.998 "name": "raid_bdev1", 00:11:55.998 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:55.998 "strip_size_kb": 0, 00:11:55.998 "state": "online", 00:11:55.998 "raid_level": "raid1", 00:11:55.998 "superblock": true, 00:11:55.998 "num_base_bdevs": 2, 00:11:55.998 "num_base_bdevs_discovered": 2, 00:11:55.998 "num_base_bdevs_operational": 2, 00:11:55.998 "process": { 00:11:55.998 "type": "rebuild", 00:11:55.999 "target": "spare", 00:11:55.999 "progress": { 00:11:55.999 "blocks": 22528, 00:11:55.999 "percent": 35 00:11:55.999 } 00:11:55.999 }, 00:11:55.999 "base_bdevs_list": [ 00:11:55.999 { 00:11:55.999 "name": "spare", 00:11:55.999 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:55.999 "is_configured": true, 00:11:55.999 "data_offset": 2048, 00:11:55.999 "data_size": 63488 00:11:55.999 }, 00:11:55.999 { 00:11:55.999 "name": "BaseBdev2", 00:11:55.999 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:55.999 "is_configured": true, 00:11:55.999 "data_offset": 2048, 00:11:55.999 "data_size": 63488 00:11:55.999 } 00:11:55.999 ] 00:11:55.999 }' 00:11:55.999 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.999 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:55.999 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.999 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:55.999 11:01:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.938 11:01:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:57.197 11:01:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.198 "name": "raid_bdev1", 00:11:57.198 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:57.198 "strip_size_kb": 0, 00:11:57.198 "state": "online", 00:11:57.198 "raid_level": "raid1", 00:11:57.198 "superblock": true, 00:11:57.198 "num_base_bdevs": 2, 00:11:57.198 "num_base_bdevs_discovered": 2, 00:11:57.198 "num_base_bdevs_operational": 2, 00:11:57.198 "process": { 00:11:57.198 "type": "rebuild", 00:11:57.198 "target": "spare", 00:11:57.198 "progress": { 00:11:57.198 "blocks": 45056, 00:11:57.198 "percent": 70 00:11:57.198 } 00:11:57.198 }, 00:11:57.198 "base_bdevs_list": [ 00:11:57.198 { 00:11:57.198 "name": "spare", 00:11:57.198 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:57.198 "is_configured": true, 00:11:57.198 "data_offset": 2048, 00:11:57.198 "data_size": 63488 00:11:57.198 }, 00:11:57.198 { 00:11:57.198 "name": "BaseBdev2", 00:11:57.198 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:57.198 "is_configured": true, 00:11:57.198 "data_offset": 2048, 00:11:57.198 "data_size": 63488 00:11:57.198 } 00:11:57.198 ] 00:11:57.198 }' 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:57.198 11:01:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:57.785 [2024-10-29 11:01:03.244014] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:57.785 [2024-10-29 11:01:03.244168] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:57.785 [2024-10-29 11:01:03.244257] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.416 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.416 "name": "raid_bdev1", 00:11:58.416 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:58.416 "strip_size_kb": 0, 00:11:58.416 "state": "online", 00:11:58.416 "raid_level": "raid1", 00:11:58.416 "superblock": true, 00:11:58.416 "num_base_bdevs": 2, 00:11:58.416 "num_base_bdevs_discovered": 2, 00:11:58.416 "num_base_bdevs_operational": 2, 00:11:58.416 "base_bdevs_list": [ 00:11:58.416 { 00:11:58.416 "name": "spare", 00:11:58.416 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:58.416 "is_configured": true, 00:11:58.416 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 }, 00:11:58.417 { 00:11:58.417 "name": "BaseBdev2", 00:11:58.417 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:58.417 "is_configured": true, 00:11:58.417 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 } 00:11:58.417 ] 00:11:58.417 }' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.417 "name": "raid_bdev1", 00:11:58.417 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:58.417 "strip_size_kb": 0, 00:11:58.417 "state": "online", 00:11:58.417 "raid_level": "raid1", 00:11:58.417 "superblock": true, 00:11:58.417 "num_base_bdevs": 2, 00:11:58.417 "num_base_bdevs_discovered": 2, 00:11:58.417 "num_base_bdevs_operational": 2, 00:11:58.417 "base_bdevs_list": [ 00:11:58.417 { 00:11:58.417 "name": "spare", 00:11:58.417 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:58.417 "is_configured": true, 00:11:58.417 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 }, 00:11:58.417 { 00:11:58.417 "name": "BaseBdev2", 00:11:58.417 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:58.417 "is_configured": true, 00:11:58.417 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 } 00:11:58.417 ] 00:11:58.417 }' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:58.417 "name": "raid_bdev1", 00:11:58.417 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:11:58.417 "strip_size_kb": 0, 00:11:58.417 "state": "online", 00:11:58.417 "raid_level": "raid1", 00:11:58.417 "superblock": true, 00:11:58.417 "num_base_bdevs": 2, 00:11:58.417 "num_base_bdevs_discovered": 2, 00:11:58.417 "num_base_bdevs_operational": 2, 00:11:58.417 "base_bdevs_list": [ 00:11:58.417 { 00:11:58.417 "name": "spare", 00:11:58.417 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:11:58.417 "is_configured": true, 00:11:58.417 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 }, 00:11:58.417 { 00:11:58.417 "name": "BaseBdev2", 00:11:58.417 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:11:58.417 "is_configured": true, 00:11:58.417 "data_offset": 2048, 00:11:58.417 "data_size": 63488 00:11:58.417 } 00:11:58.417 ] 00:11:58.417 }' 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:58.417 11:01:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.986 [2024-10-29 11:01:04.279419] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:58.986 [2024-10-29 11:01:04.279500] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:58.986 [2024-10-29 11:01:04.279628] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:58.986 [2024-10-29 11:01:04.279749] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:58.986 [2024-10-29 11:01:04.279810] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:58.986 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:59.246 /dev/nbd0 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:59.246 1+0 records in 00:11:59.246 1+0 records out 00:11:59.246 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000468785 s, 8.7 MB/s 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:59.246 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:59.504 /dev/nbd1 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:59.504 1+0 records in 00:11:59.504 1+0 records out 00:11:59.504 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000460582 s, 8.9 MB/s 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:59.504 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:59.505 11:01:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:59.763 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:59.764 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:59.764 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:59.764 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.023 [2024-10-29 11:01:05.337818] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:00.023 [2024-10-29 11:01:05.337873] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:00.023 [2024-10-29 11:01:05.337893] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:00.023 [2024-10-29 11:01:05.337906] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:00.023 [2024-10-29 11:01:05.340084] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:00.023 [2024-10-29 11:01:05.340160] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:00.023 [2024-10-29 11:01:05.340262] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:00.023 [2024-10-29 11:01:05.340317] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:00.023 [2024-10-29 11:01:05.340507] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:00.023 spare 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.023 [2024-10-29 11:01:05.440448] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:12:00.023 [2024-10-29 11:01:05.440470] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:00.023 [2024-10-29 11:01:05.440754] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1940 00:12:00.023 [2024-10-29 11:01:05.440902] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:12:00.023 [2024-10-29 11:01:05.440913] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:12:00.023 [2024-10-29 11:01:05.441028] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.023 "name": "raid_bdev1", 00:12:00.023 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:00.023 "strip_size_kb": 0, 00:12:00.023 "state": "online", 00:12:00.023 "raid_level": "raid1", 00:12:00.023 "superblock": true, 00:12:00.023 "num_base_bdevs": 2, 00:12:00.023 "num_base_bdevs_discovered": 2, 00:12:00.023 "num_base_bdevs_operational": 2, 00:12:00.023 "base_bdevs_list": [ 00:12:00.023 { 00:12:00.023 "name": "spare", 00:12:00.023 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:12:00.023 "is_configured": true, 00:12:00.023 "data_offset": 2048, 00:12:00.023 "data_size": 63488 00:12:00.023 }, 00:12:00.023 { 00:12:00.023 "name": "BaseBdev2", 00:12:00.023 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:00.023 "is_configured": true, 00:12:00.023 "data_offset": 2048, 00:12:00.023 "data_size": 63488 00:12:00.023 } 00:12:00.023 ] 00:12:00.023 }' 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.023 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:00.592 "name": "raid_bdev1", 00:12:00.592 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:00.592 "strip_size_kb": 0, 00:12:00.592 "state": "online", 00:12:00.592 "raid_level": "raid1", 00:12:00.592 "superblock": true, 00:12:00.592 "num_base_bdevs": 2, 00:12:00.592 "num_base_bdevs_discovered": 2, 00:12:00.592 "num_base_bdevs_operational": 2, 00:12:00.592 "base_bdevs_list": [ 00:12:00.592 { 00:12:00.592 "name": "spare", 00:12:00.592 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:12:00.592 "is_configured": true, 00:12:00.592 "data_offset": 2048, 00:12:00.592 "data_size": 63488 00:12:00.592 }, 00:12:00.592 { 00:12:00.592 "name": "BaseBdev2", 00:12:00.592 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:00.592 "is_configured": true, 00:12:00.592 "data_offset": 2048, 00:12:00.592 "data_size": 63488 00:12:00.592 } 00:12:00.592 ] 00:12:00.592 }' 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:00.592 11:01:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.592 [2024-10-29 11:01:06.072575] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.592 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.852 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.852 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.852 "name": "raid_bdev1", 00:12:00.852 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:00.852 "strip_size_kb": 0, 00:12:00.852 "state": "online", 00:12:00.852 "raid_level": "raid1", 00:12:00.852 "superblock": true, 00:12:00.852 "num_base_bdevs": 2, 00:12:00.852 "num_base_bdevs_discovered": 1, 00:12:00.852 "num_base_bdevs_operational": 1, 00:12:00.852 "base_bdevs_list": [ 00:12:00.852 { 00:12:00.852 "name": null, 00:12:00.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:00.852 "is_configured": false, 00:12:00.852 "data_offset": 0, 00:12:00.852 "data_size": 63488 00:12:00.852 }, 00:12:00.852 { 00:12:00.852 "name": "BaseBdev2", 00:12:00.852 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:00.852 "is_configured": true, 00:12:00.852 "data_offset": 2048, 00:12:00.852 "data_size": 63488 00:12:00.852 } 00:12:00.852 ] 00:12:00.852 }' 00:12:00.852 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.852 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.110 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:01.110 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.110 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.110 [2024-10-29 11:01:06.511835] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:01.110 [2024-10-29 11:01:06.512049] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:01.110 [2024-10-29 11:01:06.512134] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:01.110 [2024-10-29 11:01:06.512221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:01.110 [2024-10-29 11:01:06.516917] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1a10 00:12:01.110 11:01:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.110 11:01:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:01.110 [2024-10-29 11:01:06.518841] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.048 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:02.309 "name": "raid_bdev1", 00:12:02.309 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:02.309 "strip_size_kb": 0, 00:12:02.309 "state": "online", 00:12:02.309 "raid_level": "raid1", 00:12:02.309 "superblock": true, 00:12:02.309 "num_base_bdevs": 2, 00:12:02.309 "num_base_bdevs_discovered": 2, 00:12:02.309 "num_base_bdevs_operational": 2, 00:12:02.309 "process": { 00:12:02.309 "type": "rebuild", 00:12:02.309 "target": "spare", 00:12:02.309 "progress": { 00:12:02.309 "blocks": 20480, 00:12:02.309 "percent": 32 00:12:02.309 } 00:12:02.309 }, 00:12:02.309 "base_bdevs_list": [ 00:12:02.309 { 00:12:02.309 "name": "spare", 00:12:02.309 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:12:02.309 "is_configured": true, 00:12:02.309 "data_offset": 2048, 00:12:02.309 "data_size": 63488 00:12:02.309 }, 00:12:02.309 { 00:12:02.309 "name": "BaseBdev2", 00:12:02.309 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:02.309 "is_configured": true, 00:12:02.309 "data_offset": 2048, 00:12:02.309 "data_size": 63488 00:12:02.309 } 00:12:02.309 ] 00:12:02.309 }' 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.309 [2024-10-29 11:01:07.675145] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:02.309 [2024-10-29 11:01:07.722917] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:02.309 [2024-10-29 11:01:07.723053] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:02.309 [2024-10-29 11:01:07.723094] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:02.309 [2024-10-29 11:01:07.723115] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.309 "name": "raid_bdev1", 00:12:02.309 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:02.309 "strip_size_kb": 0, 00:12:02.309 "state": "online", 00:12:02.309 "raid_level": "raid1", 00:12:02.309 "superblock": true, 00:12:02.309 "num_base_bdevs": 2, 00:12:02.309 "num_base_bdevs_discovered": 1, 00:12:02.309 "num_base_bdevs_operational": 1, 00:12:02.309 "base_bdevs_list": [ 00:12:02.309 { 00:12:02.309 "name": null, 00:12:02.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:02.309 "is_configured": false, 00:12:02.309 "data_offset": 0, 00:12:02.309 "data_size": 63488 00:12:02.309 }, 00:12:02.309 { 00:12:02.309 "name": "BaseBdev2", 00:12:02.309 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:02.309 "is_configured": true, 00:12:02.309 "data_offset": 2048, 00:12:02.309 "data_size": 63488 00:12:02.309 } 00:12:02.309 ] 00:12:02.309 }' 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.309 11:01:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.878 11:01:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:02.878 11:01:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.878 11:01:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.878 [2024-10-29 11:01:08.207196] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:02.878 [2024-10-29 11:01:08.207307] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:02.878 [2024-10-29 11:01:08.207348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:12:02.878 [2024-10-29 11:01:08.207399] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:02.878 [2024-10-29 11:01:08.207875] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:02.878 [2024-10-29 11:01:08.207933] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:02.878 [2024-10-29 11:01:08.208060] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:02.878 [2024-10-29 11:01:08.208099] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:02.878 [2024-10-29 11:01:08.208163] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:02.878 [2024-10-29 11:01:08.208216] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:02.878 spare 00:12:02.878 [2024-10-29 11:01:08.212932] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:12:02.878 11:01:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.878 11:01:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:02.878 [2024-10-29 11:01:08.214795] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:03.818 "name": "raid_bdev1", 00:12:03.818 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:03.818 "strip_size_kb": 0, 00:12:03.818 "state": "online", 00:12:03.818 "raid_level": "raid1", 00:12:03.818 "superblock": true, 00:12:03.818 "num_base_bdevs": 2, 00:12:03.818 "num_base_bdevs_discovered": 2, 00:12:03.818 "num_base_bdevs_operational": 2, 00:12:03.818 "process": { 00:12:03.818 "type": "rebuild", 00:12:03.818 "target": "spare", 00:12:03.818 "progress": { 00:12:03.818 "blocks": 20480, 00:12:03.818 "percent": 32 00:12:03.818 } 00:12:03.818 }, 00:12:03.818 "base_bdevs_list": [ 00:12:03.818 { 00:12:03.818 "name": "spare", 00:12:03.818 "uuid": "ea4a434b-3128-5550-9509-80433a4f8936", 00:12:03.818 "is_configured": true, 00:12:03.818 "data_offset": 2048, 00:12:03.818 "data_size": 63488 00:12:03.818 }, 00:12:03.818 { 00:12:03.818 "name": "BaseBdev2", 00:12:03.818 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:03.818 "is_configured": true, 00:12:03.818 "data_offset": 2048, 00:12:03.818 "data_size": 63488 00:12:03.818 } 00:12:03.818 ] 00:12:03.818 }' 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:03.818 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.078 [2024-10-29 11:01:09.351743] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.078 [2024-10-29 11:01:09.418549] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:04.078 [2024-10-29 11:01:09.418668] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:04.078 [2024-10-29 11:01:09.418712] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.078 [2024-10-29 11:01:09.418749] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:04.078 "name": "raid_bdev1", 00:12:04.078 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:04.078 "strip_size_kb": 0, 00:12:04.078 "state": "online", 00:12:04.078 "raid_level": "raid1", 00:12:04.078 "superblock": true, 00:12:04.078 "num_base_bdevs": 2, 00:12:04.078 "num_base_bdevs_discovered": 1, 00:12:04.078 "num_base_bdevs_operational": 1, 00:12:04.078 "base_bdevs_list": [ 00:12:04.078 { 00:12:04.078 "name": null, 00:12:04.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.078 "is_configured": false, 00:12:04.078 "data_offset": 0, 00:12:04.078 "data_size": 63488 00:12:04.078 }, 00:12:04.078 { 00:12:04.078 "name": "BaseBdev2", 00:12:04.078 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:04.078 "is_configured": true, 00:12:04.078 "data_offset": 2048, 00:12:04.078 "data_size": 63488 00:12:04.078 } 00:12:04.078 ] 00:12:04.078 }' 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:04.078 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.653 "name": "raid_bdev1", 00:12:04.653 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:04.653 "strip_size_kb": 0, 00:12:04.653 "state": "online", 00:12:04.653 "raid_level": "raid1", 00:12:04.653 "superblock": true, 00:12:04.653 "num_base_bdevs": 2, 00:12:04.653 "num_base_bdevs_discovered": 1, 00:12:04.653 "num_base_bdevs_operational": 1, 00:12:04.653 "base_bdevs_list": [ 00:12:04.653 { 00:12:04.653 "name": null, 00:12:04.653 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.653 "is_configured": false, 00:12:04.653 "data_offset": 0, 00:12:04.653 "data_size": 63488 00:12:04.653 }, 00:12:04.653 { 00:12:04.653 "name": "BaseBdev2", 00:12:04.653 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:04.653 "is_configured": true, 00:12:04.653 "data_offset": 2048, 00:12:04.653 "data_size": 63488 00:12:04.653 } 00:12:04.653 ] 00:12:04.653 }' 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:04.653 11:01:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:04.653 [2024-10-29 11:01:10.022411] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:04.653 [2024-10-29 11:01:10.022521] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:04.653 [2024-10-29 11:01:10.022546] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:04.653 [2024-10-29 11:01:10.022558] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:04.653 [2024-10-29 11:01:10.022960] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:04.653 [2024-10-29 11:01:10.022981] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:04.653 [2024-10-29 11:01:10.023054] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:04.653 [2024-10-29 11:01:10.023072] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:04.653 [2024-10-29 11:01:10.023090] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:04.653 [2024-10-29 11:01:10.023102] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:04.653 BaseBdev1 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.653 11:01:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.592 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.593 "name": "raid_bdev1", 00:12:05.593 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:05.593 "strip_size_kb": 0, 00:12:05.593 "state": "online", 00:12:05.593 "raid_level": "raid1", 00:12:05.593 "superblock": true, 00:12:05.593 "num_base_bdevs": 2, 00:12:05.593 "num_base_bdevs_discovered": 1, 00:12:05.593 "num_base_bdevs_operational": 1, 00:12:05.593 "base_bdevs_list": [ 00:12:05.593 { 00:12:05.593 "name": null, 00:12:05.593 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.593 "is_configured": false, 00:12:05.593 "data_offset": 0, 00:12:05.593 "data_size": 63488 00:12:05.593 }, 00:12:05.593 { 00:12:05.593 "name": "BaseBdev2", 00:12:05.593 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:05.593 "is_configured": true, 00:12:05.593 "data_offset": 2048, 00:12:05.593 "data_size": 63488 00:12:05.593 } 00:12:05.593 ] 00:12:05.593 }' 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.593 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.159 "name": "raid_bdev1", 00:12:06.159 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:06.159 "strip_size_kb": 0, 00:12:06.159 "state": "online", 00:12:06.159 "raid_level": "raid1", 00:12:06.159 "superblock": true, 00:12:06.159 "num_base_bdevs": 2, 00:12:06.159 "num_base_bdevs_discovered": 1, 00:12:06.159 "num_base_bdevs_operational": 1, 00:12:06.159 "base_bdevs_list": [ 00:12:06.159 { 00:12:06.159 "name": null, 00:12:06.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.159 "is_configured": false, 00:12:06.159 "data_offset": 0, 00:12:06.159 "data_size": 63488 00:12:06.159 }, 00:12:06.159 { 00:12:06.159 "name": "BaseBdev2", 00:12:06.159 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:06.159 "is_configured": true, 00:12:06.159 "data_offset": 2048, 00:12:06.159 "data_size": 63488 00:12:06.159 } 00:12:06.159 ] 00:12:06.159 }' 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.159 [2024-10-29 11:01:11.627746] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:06.159 [2024-10-29 11:01:11.627961] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:06.159 [2024-10-29 11:01:11.628017] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:06.159 request: 00:12:06.159 { 00:12:06.159 "base_bdev": "BaseBdev1", 00:12:06.159 "raid_bdev": "raid_bdev1", 00:12:06.159 "method": "bdev_raid_add_base_bdev", 00:12:06.159 "req_id": 1 00:12:06.159 } 00:12:06.159 Got JSON-RPC error response 00:12:06.159 response: 00:12:06.159 { 00:12:06.159 "code": -22, 00:12:06.159 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:06.159 } 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:06.159 11:01:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:07.535 "name": "raid_bdev1", 00:12:07.535 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:07.535 "strip_size_kb": 0, 00:12:07.535 "state": "online", 00:12:07.535 "raid_level": "raid1", 00:12:07.535 "superblock": true, 00:12:07.535 "num_base_bdevs": 2, 00:12:07.535 "num_base_bdevs_discovered": 1, 00:12:07.535 "num_base_bdevs_operational": 1, 00:12:07.535 "base_bdevs_list": [ 00:12:07.535 { 00:12:07.535 "name": null, 00:12:07.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.535 "is_configured": false, 00:12:07.535 "data_offset": 0, 00:12:07.535 "data_size": 63488 00:12:07.535 }, 00:12:07.535 { 00:12:07.535 "name": "BaseBdev2", 00:12:07.535 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:07.535 "is_configured": true, 00:12:07.535 "data_offset": 2048, 00:12:07.535 "data_size": 63488 00:12:07.535 } 00:12:07.535 ] 00:12:07.535 }' 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:07.535 11:01:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.796 "name": "raid_bdev1", 00:12:07.796 "uuid": "74e59812-076a-44c6-a5f8-c3589acb41bf", 00:12:07.796 "strip_size_kb": 0, 00:12:07.796 "state": "online", 00:12:07.796 "raid_level": "raid1", 00:12:07.796 "superblock": true, 00:12:07.796 "num_base_bdevs": 2, 00:12:07.796 "num_base_bdevs_discovered": 1, 00:12:07.796 "num_base_bdevs_operational": 1, 00:12:07.796 "base_bdevs_list": [ 00:12:07.796 { 00:12:07.796 "name": null, 00:12:07.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.796 "is_configured": false, 00:12:07.796 "data_offset": 0, 00:12:07.796 "data_size": 63488 00:12:07.796 }, 00:12:07.796 { 00:12:07.796 "name": "BaseBdev2", 00:12:07.796 "uuid": "4353e073-0302-502f-9e8e-bfa7d5fb3170", 00:12:07.796 "is_configured": true, 00:12:07.796 "data_offset": 2048, 00:12:07.796 "data_size": 63488 00:12:07.796 } 00:12:07.796 ] 00:12:07.796 }' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86572 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@952 -- # '[' -z 86572 ']' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # kill -0 86572 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@957 -- # uname 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 86572 00:12:07.796 killing process with pid 86572 00:12:07.796 Received shutdown signal, test time was about 60.000000 seconds 00:12:07.796 00:12:07.796 Latency(us) 00:12:07.796 [2024-10-29T11:01:13.294Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:07.796 [2024-10-29T11:01:13.294Z] =================================================================================================================== 00:12:07.796 [2024-10-29T11:01:13.294Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 86572' 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@971 -- # kill 86572 00:12:07.796 [2024-10-29 11:01:13.272270] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:07.796 [2024-10-29 11:01:13.272402] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:07.796 [2024-10-29 11:01:13.272454] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:07.796 [2024-10-29 11:01:13.272463] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:12:07.796 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@976 -- # wait 86572 00:12:08.055 [2024-10-29 11:01:13.303521] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:08.055 11:01:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:08.055 ************************************ 00:12:08.056 END TEST raid_rebuild_test_sb 00:12:08.056 ************************************ 00:12:08.056 00:12:08.056 real 0m21.255s 00:12:08.056 user 0m26.497s 00:12:08.056 sys 0m3.405s 00:12:08.056 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:12:08.056 11:01:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 11:01:13 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:12:08.314 11:01:13 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:12:08.314 11:01:13 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:12:08.314 11:01:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:08.314 ************************************ 00:12:08.314 START TEST raid_rebuild_test_io 00:12:08.314 ************************************ 00:12:08.314 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 false true true 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87280 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87280 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@833 -- # '[' -z 87280 ']' 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # local max_retries=100 00:12:08.315 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # xtrace_disable 00:12:08.315 11:01:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.315 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:08.315 Zero copy mechanism will not be used. 00:12:08.315 [2024-10-29 11:01:13.678825] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:12:08.315 [2024-10-29 11:01:13.678942] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87280 ] 00:12:08.574 [2024-10-29 11:01:13.827877] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:08.574 [2024-10-29 11:01:13.851931] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:08.574 [2024-10-29 11:01:13.892913] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:08.574 [2024-10-29 11:01:13.892945] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # return 0 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 BaseBdev1_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 [2024-10-29 11:01:14.522439] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:09.144 [2024-10-29 11:01:14.522506] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.144 [2024-10-29 11:01:14.522548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:09.144 [2024-10-29 11:01:14.522562] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.144 [2024-10-29 11:01:14.524602] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.144 [2024-10-29 11:01:14.524710] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:09.144 BaseBdev1 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 BaseBdev2_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 [2024-10-29 11:01:14.550690] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:09.144 [2024-10-29 11:01:14.550746] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.144 [2024-10-29 11:01:14.550781] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:09.144 [2024-10-29 11:01:14.550790] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.144 [2024-10-29 11:01:14.552803] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.144 [2024-10-29 11:01:14.552840] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:09.144 BaseBdev2 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 spare_malloc 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 spare_delay 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 [2024-10-29 11:01:14.590971] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:09.144 [2024-10-29 11:01:14.591106] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:09.144 [2024-10-29 11:01:14.591130] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:09.144 [2024-10-29 11:01:14.591138] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:09.144 [2024-10-29 11:01:14.593177] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:09.144 [2024-10-29 11:01:14.593212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:09.144 spare 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.144 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.144 [2024-10-29 11:01:14.602986] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:09.144 [2024-10-29 11:01:14.604752] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:09.144 [2024-10-29 11:01:14.604899] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:09.144 [2024-10-29 11:01:14.604914] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:09.144 [2024-10-29 11:01:14.605161] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:12:09.144 [2024-10-29 11:01:14.605285] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:09.145 [2024-10-29 11:01:14.605297] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:09.145 [2024-10-29 11:01:14.605427] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.145 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.404 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.404 "name": "raid_bdev1", 00:12:09.404 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:09.404 "strip_size_kb": 0, 00:12:09.404 "state": "online", 00:12:09.404 "raid_level": "raid1", 00:12:09.404 "superblock": false, 00:12:09.404 "num_base_bdevs": 2, 00:12:09.404 "num_base_bdevs_discovered": 2, 00:12:09.404 "num_base_bdevs_operational": 2, 00:12:09.404 "base_bdevs_list": [ 00:12:09.404 { 00:12:09.404 "name": "BaseBdev1", 00:12:09.404 "uuid": "1da9f87f-de8f-50be-8086-d8e10c6daee8", 00:12:09.404 "is_configured": true, 00:12:09.404 "data_offset": 0, 00:12:09.404 "data_size": 65536 00:12:09.404 }, 00:12:09.404 { 00:12:09.404 "name": "BaseBdev2", 00:12:09.404 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:09.404 "is_configured": true, 00:12:09.404 "data_offset": 0, 00:12:09.404 "data_size": 65536 00:12:09.404 } 00:12:09.404 ] 00:12:09.404 }' 00:12:09.404 11:01:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.404 11:01:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.674 [2024-10-29 11:01:15.062468] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.674 [2024-10-29 11:01:15.138037] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.674 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.948 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.948 "name": "raid_bdev1", 00:12:09.948 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:09.948 "strip_size_kb": 0, 00:12:09.948 "state": "online", 00:12:09.948 "raid_level": "raid1", 00:12:09.948 "superblock": false, 00:12:09.948 "num_base_bdevs": 2, 00:12:09.948 "num_base_bdevs_discovered": 1, 00:12:09.948 "num_base_bdevs_operational": 1, 00:12:09.948 "base_bdevs_list": [ 00:12:09.948 { 00:12:09.948 "name": null, 00:12:09.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.948 "is_configured": false, 00:12:09.948 "data_offset": 0, 00:12:09.948 "data_size": 65536 00:12:09.948 }, 00:12:09.948 { 00:12:09.948 "name": "BaseBdev2", 00:12:09.948 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:09.948 "is_configured": true, 00:12:09.948 "data_offset": 0, 00:12:09.948 "data_size": 65536 00:12:09.948 } 00:12:09.948 ] 00:12:09.948 }' 00:12:09.948 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.948 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.948 [2024-10-29 11:01:15.223830] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:12:09.948 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:09.948 Zero copy mechanism will not be used. 00:12:09.948 Running I/O for 60 seconds... 00:12:10.207 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:10.207 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.207 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.207 [2024-10-29 11:01:15.541525] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:10.207 11:01:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.207 11:01:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:10.207 [2024-10-29 11:01:15.595520] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:10.207 [2024-10-29 11:01:15.597493] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:10.207 [2024-10-29 11:01:15.704338] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:10.207 [2024-10-29 11:01:15.704673] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:10.466 [2024-10-29 11:01:15.813149] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:10.466 [2024-10-29 11:01:15.813459] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:10.725 [2024-10-29 11:01:16.025807] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:10.725 [2024-10-29 11:01:16.026170] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:10.725 [2024-10-29 11:01:16.150266] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:10.725 [2024-10-29 11:01:16.150525] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:11.243 137.00 IOPS, 411.00 MiB/s [2024-10-29T11:01:16.741Z] 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:11.243 [2024-10-29 11:01:16.584080] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.243 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:11.244 "name": "raid_bdev1", 00:12:11.244 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:11.244 "strip_size_kb": 0, 00:12:11.244 "state": "online", 00:12:11.244 "raid_level": "raid1", 00:12:11.244 "superblock": false, 00:12:11.244 "num_base_bdevs": 2, 00:12:11.244 "num_base_bdevs_discovered": 2, 00:12:11.244 "num_base_bdevs_operational": 2, 00:12:11.244 "process": { 00:12:11.244 "type": "rebuild", 00:12:11.244 "target": "spare", 00:12:11.244 "progress": { 00:12:11.244 "blocks": 16384, 00:12:11.244 "percent": 25 00:12:11.244 } 00:12:11.244 }, 00:12:11.244 "base_bdevs_list": [ 00:12:11.244 { 00:12:11.244 "name": "spare", 00:12:11.244 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:11.244 "is_configured": true, 00:12:11.244 "data_offset": 0, 00:12:11.244 "data_size": 65536 00:12:11.244 }, 00:12:11.244 { 00:12:11.244 "name": "BaseBdev2", 00:12:11.244 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:11.244 "is_configured": true, 00:12:11.244 "data_offset": 0, 00:12:11.244 "data_size": 65536 00:12:11.244 } 00:12:11.244 ] 00:12:11.244 }' 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.244 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.244 [2024-10-29 11:01:16.730241] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:11.503 [2024-10-29 11:01:16.900009] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:11.503 [2024-10-29 11:01:16.913178] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:11.503 [2024-10-29 11:01:16.913218] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:11.503 [2024-10-29 11:01:16.913233] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:11.503 [2024-10-29 11:01:16.930034] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.503 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.504 "name": "raid_bdev1", 00:12:11.504 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:11.504 "strip_size_kb": 0, 00:12:11.504 "state": "online", 00:12:11.504 "raid_level": "raid1", 00:12:11.504 "superblock": false, 00:12:11.504 "num_base_bdevs": 2, 00:12:11.504 "num_base_bdevs_discovered": 1, 00:12:11.504 "num_base_bdevs_operational": 1, 00:12:11.504 "base_bdevs_list": [ 00:12:11.504 { 00:12:11.504 "name": null, 00:12:11.504 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:11.504 "is_configured": false, 00:12:11.504 "data_offset": 0, 00:12:11.504 "data_size": 65536 00:12:11.504 }, 00:12:11.504 { 00:12:11.504 "name": "BaseBdev2", 00:12:11.504 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:11.504 "is_configured": true, 00:12:11.504 "data_offset": 0, 00:12:11.504 "data_size": 65536 00:12:11.504 } 00:12:11.504 ] 00:12:11.504 }' 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.504 11:01:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.022 153.00 IOPS, 459.00 MiB/s [2024-10-29T11:01:17.520Z] 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:12.022 "name": "raid_bdev1", 00:12:12.022 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:12.022 "strip_size_kb": 0, 00:12:12.022 "state": "online", 00:12:12.022 "raid_level": "raid1", 00:12:12.022 "superblock": false, 00:12:12.022 "num_base_bdevs": 2, 00:12:12.022 "num_base_bdevs_discovered": 1, 00:12:12.022 "num_base_bdevs_operational": 1, 00:12:12.022 "base_bdevs_list": [ 00:12:12.022 { 00:12:12.022 "name": null, 00:12:12.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:12.022 "is_configured": false, 00:12:12.022 "data_offset": 0, 00:12:12.022 "data_size": 65536 00:12:12.022 }, 00:12:12.022 { 00:12:12.022 "name": "BaseBdev2", 00:12:12.022 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:12.022 "is_configured": true, 00:12:12.022 "data_offset": 0, 00:12:12.022 "data_size": 65536 00:12:12.022 } 00:12:12.022 ] 00:12:12.022 }' 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:12.022 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:12.023 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.023 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.282 [2024-10-29 11:01:17.532146] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:12.282 11:01:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.282 11:01:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:12.282 [2024-10-29 11:01:17.597071] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:12.282 [2024-10-29 11:01:17.599071] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:12.282 [2024-10-29 11:01:17.712056] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:12.282 [2024-10-29 11:01:17.712416] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:12.541 [2024-10-29 11:01:17.826490] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:12.541 [2024-10-29 11:01:17.826721] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:12.801 [2024-10-29 11:01:18.161490] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:13.061 173.00 IOPS, 519.00 MiB/s [2024-10-29T11:01:18.559Z] [2024-10-29 11:01:18.369847] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:13.061 [2024-10-29 11:01:18.370121] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.320 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.321 [2024-10-29 11:01:18.589314] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.321 "name": "raid_bdev1", 00:12:13.321 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:13.321 "strip_size_kb": 0, 00:12:13.321 "state": "online", 00:12:13.321 "raid_level": "raid1", 00:12:13.321 "superblock": false, 00:12:13.321 "num_base_bdevs": 2, 00:12:13.321 "num_base_bdevs_discovered": 2, 00:12:13.321 "num_base_bdevs_operational": 2, 00:12:13.321 "process": { 00:12:13.321 "type": "rebuild", 00:12:13.321 "target": "spare", 00:12:13.321 "progress": { 00:12:13.321 "blocks": 12288, 00:12:13.321 "percent": 18 00:12:13.321 } 00:12:13.321 }, 00:12:13.321 "base_bdevs_list": [ 00:12:13.321 { 00:12:13.321 "name": "spare", 00:12:13.321 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:13.321 "is_configured": true, 00:12:13.321 "data_offset": 0, 00:12:13.321 "data_size": 65536 00:12:13.321 }, 00:12:13.321 { 00:12:13.321 "name": "BaseBdev2", 00:12:13.321 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:13.321 "is_configured": true, 00:12:13.321 "data_offset": 0, 00:12:13.321 "data_size": 65536 00:12:13.321 } 00:12:13.321 ] 00:12:13.321 }' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=323 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.321 "name": "raid_bdev1", 00:12:13.321 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:13.321 "strip_size_kb": 0, 00:12:13.321 "state": "online", 00:12:13.321 "raid_level": "raid1", 00:12:13.321 "superblock": false, 00:12:13.321 "num_base_bdevs": 2, 00:12:13.321 "num_base_bdevs_discovered": 2, 00:12:13.321 "num_base_bdevs_operational": 2, 00:12:13.321 "process": { 00:12:13.321 "type": "rebuild", 00:12:13.321 "target": "spare", 00:12:13.321 "progress": { 00:12:13.321 "blocks": 14336, 00:12:13.321 "percent": 21 00:12:13.321 } 00:12:13.321 }, 00:12:13.321 "base_bdevs_list": [ 00:12:13.321 { 00:12:13.321 "name": "spare", 00:12:13.321 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:13.321 "is_configured": true, 00:12:13.321 "data_offset": 0, 00:12:13.321 "data_size": 65536 00:12:13.321 }, 00:12:13.321 { 00:12:13.321 "name": "BaseBdev2", 00:12:13.321 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:13.321 "is_configured": true, 00:12:13.321 "data_offset": 0, 00:12:13.321 "data_size": 65536 00:12:13.321 } 00:12:13.321 ] 00:12:13.321 }' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:13.321 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:13.321 [2024-10-29 11:01:18.806347] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:13.321 [2024-10-29 11:01:18.806687] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:13.581 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:13.581 11:01:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:13.840 149.25 IOPS, 447.75 MiB/s [2024-10-29T11:01:19.338Z] [2024-10-29 11:01:19.251662] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:14.099 [2024-10-29 11:01:19.587062] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.358 11:01:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:14.617 "name": "raid_bdev1", 00:12:14.617 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:14.617 "strip_size_kb": 0, 00:12:14.617 "state": "online", 00:12:14.617 "raid_level": "raid1", 00:12:14.617 "superblock": false, 00:12:14.617 "num_base_bdevs": 2, 00:12:14.617 "num_base_bdevs_discovered": 2, 00:12:14.617 "num_base_bdevs_operational": 2, 00:12:14.617 "process": { 00:12:14.617 "type": "rebuild", 00:12:14.617 "target": "spare", 00:12:14.617 "progress": { 00:12:14.617 "blocks": 30720, 00:12:14.617 "percent": 46 00:12:14.617 } 00:12:14.617 }, 00:12:14.617 "base_bdevs_list": [ 00:12:14.617 { 00:12:14.617 "name": "spare", 00:12:14.617 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:14.617 "is_configured": true, 00:12:14.617 "data_offset": 0, 00:12:14.617 "data_size": 65536 00:12:14.617 }, 00:12:14.617 { 00:12:14.617 "name": "BaseBdev2", 00:12:14.617 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:14.617 "is_configured": true, 00:12:14.617 "data_offset": 0, 00:12:14.617 "data_size": 65536 00:12:14.617 } 00:12:14.617 ] 00:12:14.617 }' 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:14.617 [2024-10-29 11:01:19.920018] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:14.617 11:01:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:14.875 [2024-10-29 11:01:20.132339] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:14.875 [2024-10-29 11:01:20.132699] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:15.811 132.80 IOPS, 398.40 MiB/s [2024-10-29T11:01:21.309Z] 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.811 11:01:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:15.811 "name": "raid_bdev1", 00:12:15.811 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:15.811 "strip_size_kb": 0, 00:12:15.811 "state": "online", 00:12:15.811 "raid_level": "raid1", 00:12:15.811 "superblock": false, 00:12:15.811 "num_base_bdevs": 2, 00:12:15.811 "num_base_bdevs_discovered": 2, 00:12:15.811 "num_base_bdevs_operational": 2, 00:12:15.811 "process": { 00:12:15.811 "type": "rebuild", 00:12:15.811 "target": "spare", 00:12:15.811 "progress": { 00:12:15.811 "blocks": 49152, 00:12:15.811 "percent": 75 00:12:15.811 } 00:12:15.811 }, 00:12:15.811 "base_bdevs_list": [ 00:12:15.811 { 00:12:15.811 "name": "spare", 00:12:15.811 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:15.811 "is_configured": true, 00:12:15.811 "data_offset": 0, 00:12:15.811 "data_size": 65536 00:12:15.811 }, 00:12:15.811 { 00:12:15.811 "name": "BaseBdev2", 00:12:15.811 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:15.811 "is_configured": true, 00:12:15.811 "data_offset": 0, 00:12:15.811 "data_size": 65536 00:12:15.811 } 00:12:15.811 ] 00:12:15.811 }' 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:15.811 [2024-10-29 11:01:21.120317] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:15.811 11:01:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:16.071 116.50 IOPS, 349.50 MiB/s [2024-10-29T11:01:21.569Z] [2024-10-29 11:01:21.445886] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:12:16.639 [2024-10-29 11:01:21.980945] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:16.640 [2024-10-29 11:01:22.086116] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:16.640 [2024-10-29 11:01:22.087984] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.640 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:16.899 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.899 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:16.899 "name": "raid_bdev1", 00:12:16.899 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:16.899 "strip_size_kb": 0, 00:12:16.899 "state": "online", 00:12:16.899 "raid_level": "raid1", 00:12:16.899 "superblock": false, 00:12:16.899 "num_base_bdevs": 2, 00:12:16.899 "num_base_bdevs_discovered": 2, 00:12:16.899 "num_base_bdevs_operational": 2, 00:12:16.899 "base_bdevs_list": [ 00:12:16.899 { 00:12:16.899 "name": "spare", 00:12:16.899 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:16.899 "is_configured": true, 00:12:16.899 "data_offset": 0, 00:12:16.899 "data_size": 65536 00:12:16.899 }, 00:12:16.899 { 00:12:16.900 "name": "BaseBdev2", 00:12:16.900 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:16.900 "is_configured": true, 00:12:16.900 "data_offset": 0, 00:12:16.900 "data_size": 65536 00:12:16.900 } 00:12:16.900 ] 00:12:16.900 }' 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:16.900 104.86 IOPS, 314.57 MiB/s [2024-10-29T11:01:22.398Z] 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:16.900 "name": "raid_bdev1", 00:12:16.900 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:16.900 "strip_size_kb": 0, 00:12:16.900 "state": "online", 00:12:16.900 "raid_level": "raid1", 00:12:16.900 "superblock": false, 00:12:16.900 "num_base_bdevs": 2, 00:12:16.900 "num_base_bdevs_discovered": 2, 00:12:16.900 "num_base_bdevs_operational": 2, 00:12:16.900 "base_bdevs_list": [ 00:12:16.900 { 00:12:16.900 "name": "spare", 00:12:16.900 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:16.900 "is_configured": true, 00:12:16.900 "data_offset": 0, 00:12:16.900 "data_size": 65536 00:12:16.900 }, 00:12:16.900 { 00:12:16.900 "name": "BaseBdev2", 00:12:16.900 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:16.900 "is_configured": true, 00:12:16.900 "data_offset": 0, 00:12:16.900 "data_size": 65536 00:12:16.900 } 00:12:16.900 ] 00:12:16.900 }' 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:16.900 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:17.160 "name": "raid_bdev1", 00:12:17.160 "uuid": "d871b08c-c183-427f-b042-05addcdb3773", 00:12:17.160 "strip_size_kb": 0, 00:12:17.160 "state": "online", 00:12:17.160 "raid_level": "raid1", 00:12:17.160 "superblock": false, 00:12:17.160 "num_base_bdevs": 2, 00:12:17.160 "num_base_bdevs_discovered": 2, 00:12:17.160 "num_base_bdevs_operational": 2, 00:12:17.160 "base_bdevs_list": [ 00:12:17.160 { 00:12:17.160 "name": "spare", 00:12:17.160 "uuid": "6918fb8c-e06b-5178-9304-56a3b12a28ae", 00:12:17.160 "is_configured": true, 00:12:17.160 "data_offset": 0, 00:12:17.160 "data_size": 65536 00:12:17.160 }, 00:12:17.160 { 00:12:17.160 "name": "BaseBdev2", 00:12:17.160 "uuid": "90b10bc2-898e-5242-8bb6-5bb860bdf485", 00:12:17.160 "is_configured": true, 00:12:17.160 "data_offset": 0, 00:12:17.160 "data_size": 65536 00:12:17.160 } 00:12:17.160 ] 00:12:17.160 }' 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:17.160 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.420 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:17.420 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.420 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.420 [2024-10-29 11:01:22.894264] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:17.420 [2024-10-29 11:01:22.894386] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:17.681 00:12:17.681 Latency(us) 00:12:17.681 [2024-10-29T11:01:23.179Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:17.681 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:17.681 raid_bdev1 : 7.74 98.33 295.00 0.00 0.00 14071.04 255.78 111726.00 00:12:17.681 [2024-10-29T11:01:23.179Z] =================================================================================================================== 00:12:17.681 [2024-10-29T11:01:23.179Z] Total : 98.33 295.00 0.00 0.00 14071.04 255.78 111726.00 00:12:17.681 [2024-10-29 11:01:22.953745] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.681 [2024-10-29 11:01:22.953854] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:17.681 [2024-10-29 11:01:22.953980] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:17.681 [2024-10-29 11:01:22.954045] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:17.681 { 00:12:17.681 "results": [ 00:12:17.681 { 00:12:17.681 "job": "raid_bdev1", 00:12:17.681 "core_mask": "0x1", 00:12:17.681 "workload": "randrw", 00:12:17.681 "percentage": 50, 00:12:17.681 "status": "finished", 00:12:17.681 "queue_depth": 2, 00:12:17.681 "io_size": 3145728, 00:12:17.681 "runtime": 7.739103, 00:12:17.681 "iops": 98.33180925489685, 00:12:17.681 "mibps": 294.9954277646906, 00:12:17.681 "io_failed": 0, 00:12:17.681 "io_timeout": 0, 00:12:17.681 "avg_latency_us": 14071.038259242896, 00:12:17.681 "min_latency_us": 255.7764192139738, 00:12:17.681 "max_latency_us": 111726.00174672488 00:12:17.681 } 00:12:17.681 ], 00:12:17.681 "core_count": 1 00:12:17.681 } 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:17.681 11:01:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:17.681 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:17.942 /dev/nbd0 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # local i 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # break 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:17.942 1+0 records in 00:12:17.942 1+0 records out 00:12:17.942 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000409891 s, 10.0 MB/s 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # size=4096 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # return 0 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:17.942 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:18.202 /dev/nbd1 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # local i 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:12:18.202 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # break 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:18.203 1+0 records in 00:12:18.203 1+0 records out 00:12:18.203 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000372787 s, 11.0 MB/s 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # size=4096 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # return 0 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:18.203 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:18.463 11:01:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 87280 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@952 -- # '[' -z 87280 ']' 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # kill -0 87280 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@957 -- # uname 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 87280 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@970 -- # echo 'killing process with pid 87280' 00:12:18.723 killing process with pid 87280 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@971 -- # kill 87280 00:12:18.723 Received shutdown signal, test time was about 8.850990 seconds 00:12:18.723 00:12:18.723 Latency(us) 00:12:18.723 [2024-10-29T11:01:24.221Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:18.723 [2024-10-29T11:01:24.221Z] =================================================================================================================== 00:12:18.723 [2024-10-29T11:01:24.221Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:18.723 [2024-10-29 11:01:24.059854] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:18.723 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@976 -- # wait 87280 00:12:18.723 [2024-10-29 11:01:24.086245] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:18.984 00:12:18.984 real 0m10.701s 00:12:18.984 user 0m13.830s 00:12:18.984 sys 0m1.372s 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1128 -- # xtrace_disable 00:12:18.984 ************************************ 00:12:18.984 END TEST raid_rebuild_test_io 00:12:18.984 ************************************ 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.984 11:01:24 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:12:18.984 11:01:24 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:12:18.984 11:01:24 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:12:18.984 11:01:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:18.984 ************************************ 00:12:18.984 START TEST raid_rebuild_test_sb_io 00:12:18.984 ************************************ 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 true true true 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87643 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87643 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@833 -- # '[' -z 87643 ']' 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # local max_retries=100 00:12:18.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # xtrace_disable 00:12:18.984 11:01:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.984 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:18.984 Zero copy mechanism will not be used. 00:12:18.984 [2024-10-29 11:01:24.469071] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:12:18.984 [2024-10-29 11:01:24.469212] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87643 ] 00:12:19.244 [2024-10-29 11:01:24.641973] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:19.244 [2024-10-29 11:01:24.667462] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:19.244 [2024-10-29 11:01:24.710759] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:19.244 [2024-10-29 11:01:24.710879] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # return 0 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:19.813 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.073 BaseBdev1_malloc 00:12:20.073 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 [2024-10-29 11:01:25.332811] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:20.074 [2024-10-29 11:01:25.332979] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.074 [2024-10-29 11:01:25.333014] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:20.074 [2024-10-29 11:01:25.333029] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.074 [2024-10-29 11:01:25.335285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.074 [2024-10-29 11:01:25.335321] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:20.074 BaseBdev1 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 BaseBdev2_malloc 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 [2024-10-29 11:01:25.361254] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:20.074 [2024-10-29 11:01:25.361318] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.074 [2024-10-29 11:01:25.361357] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:20.074 [2024-10-29 11:01:25.361366] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.074 [2024-10-29 11:01:25.363488] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.074 [2024-10-29 11:01:25.363562] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:20.074 BaseBdev2 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 spare_malloc 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 spare_delay 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 [2024-10-29 11:01:25.401879] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:20.074 [2024-10-29 11:01:25.401947] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:20.074 [2024-10-29 11:01:25.401970] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:12:20.074 [2024-10-29 11:01:25.401979] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:20.074 [2024-10-29 11:01:25.404316] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:20.074 [2024-10-29 11:01:25.404354] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:20.074 spare 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 [2024-10-29 11:01:25.413896] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:20.074 [2024-10-29 11:01:25.415802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:20.074 [2024-10-29 11:01:25.415957] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:20.074 [2024-10-29 11:01:25.415970] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:20.074 [2024-10-29 11:01:25.416208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:12:20.074 [2024-10-29 11:01:25.416344] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:20.074 [2024-10-29 11:01:25.416378] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:20.074 [2024-10-29 11:01:25.416502] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.074 "name": "raid_bdev1", 00:12:20.074 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:20.074 "strip_size_kb": 0, 00:12:20.074 "state": "online", 00:12:20.074 "raid_level": "raid1", 00:12:20.074 "superblock": true, 00:12:20.074 "num_base_bdevs": 2, 00:12:20.074 "num_base_bdevs_discovered": 2, 00:12:20.074 "num_base_bdevs_operational": 2, 00:12:20.074 "base_bdevs_list": [ 00:12:20.074 { 00:12:20.074 "name": "BaseBdev1", 00:12:20.074 "uuid": "57c69bb7-c855-5091-a82d-57814f978acb", 00:12:20.074 "is_configured": true, 00:12:20.074 "data_offset": 2048, 00:12:20.074 "data_size": 63488 00:12:20.074 }, 00:12:20.074 { 00:12:20.074 "name": "BaseBdev2", 00:12:20.074 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:20.074 "is_configured": true, 00:12:20.074 "data_offset": 2048, 00:12:20.074 "data_size": 63488 00:12:20.074 } 00:12:20.074 ] 00:12:20.074 }' 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.074 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 [2024-10-29 11:01:25.857426] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 [2024-10-29 11:01:25.956939] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 11:01:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.649 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.649 "name": "raid_bdev1", 00:12:20.649 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:20.649 "strip_size_kb": 0, 00:12:20.649 "state": "online", 00:12:20.649 "raid_level": "raid1", 00:12:20.649 "superblock": true, 00:12:20.649 "num_base_bdevs": 2, 00:12:20.649 "num_base_bdevs_discovered": 1, 00:12:20.649 "num_base_bdevs_operational": 1, 00:12:20.649 "base_bdevs_list": [ 00:12:20.649 { 00:12:20.649 "name": null, 00:12:20.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.649 "is_configured": false, 00:12:20.649 "data_offset": 0, 00:12:20.649 "data_size": 63488 00:12:20.649 }, 00:12:20.649 { 00:12:20.649 "name": "BaseBdev2", 00:12:20.649 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:20.649 "is_configured": true, 00:12:20.649 "data_offset": 2048, 00:12:20.649 "data_size": 63488 00:12:20.649 } 00:12:20.649 ] 00:12:20.649 }' 00:12:20.649 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.649 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.649 [2024-10-29 11:01:26.050751] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:12:20.649 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:20.649 Zero copy mechanism will not be used. 00:12:20.649 Running I/O for 60 seconds... 00:12:20.909 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:20.910 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.910 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.910 [2024-10-29 11:01:26.393995] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:21.170 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.170 11:01:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:21.170 [2024-10-29 11:01:26.452158] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:21.170 [2024-10-29 11:01:26.454137] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:21.170 [2024-10-29 11:01:26.567583] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:21.170 [2024-10-29 11:01:26.568144] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:21.429 [2024-10-29 11:01:26.795166] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:21.429 [2024-10-29 11:01:26.795475] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:21.949 170.00 IOPS, 510.00 MiB/s [2024-10-29T11:01:27.447Z] [2024-10-29 11:01:27.238728] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.208 [2024-10-29 11:01:27.459228] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:22.208 [2024-10-29 11:01:27.459647] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.208 "name": "raid_bdev1", 00:12:22.208 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:22.208 "strip_size_kb": 0, 00:12:22.208 "state": "online", 00:12:22.208 "raid_level": "raid1", 00:12:22.208 "superblock": true, 00:12:22.208 "num_base_bdevs": 2, 00:12:22.208 "num_base_bdevs_discovered": 2, 00:12:22.208 "num_base_bdevs_operational": 2, 00:12:22.208 "process": { 00:12:22.208 "type": "rebuild", 00:12:22.208 "target": "spare", 00:12:22.208 "progress": { 00:12:22.208 "blocks": 12288, 00:12:22.208 "percent": 19 00:12:22.208 } 00:12:22.208 }, 00:12:22.208 "base_bdevs_list": [ 00:12:22.208 { 00:12:22.208 "name": "spare", 00:12:22.208 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:22.208 "is_configured": true, 00:12:22.208 "data_offset": 2048, 00:12:22.208 "data_size": 63488 00:12:22.208 }, 00:12:22.208 { 00:12:22.208 "name": "BaseBdev2", 00:12:22.208 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:22.208 "is_configured": true, 00:12:22.208 "data_offset": 2048, 00:12:22.208 "data_size": 63488 00:12:22.208 } 00:12:22.208 ] 00:12:22.208 }' 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.208 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.209 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.209 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.209 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:22.209 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.209 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.209 [2024-10-29 11:01:27.579749] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:22.209 [2024-10-29 11:01:27.672021] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:22.209 [2024-10-29 11:01:27.672325] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:22.209 [2024-10-29 11:01:27.673334] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:22.209 [2024-10-29 11:01:27.680785] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:22.209 [2024-10-29 11:01:27.680815] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:22.209 [2024-10-29 11:01:27.680828] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:22.209 [2024-10-29 11:01:27.702257] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.468 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:22.468 "name": "raid_bdev1", 00:12:22.468 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:22.468 "strip_size_kb": 0, 00:12:22.468 "state": "online", 00:12:22.468 "raid_level": "raid1", 00:12:22.468 "superblock": true, 00:12:22.468 "num_base_bdevs": 2, 00:12:22.468 "num_base_bdevs_discovered": 1, 00:12:22.468 "num_base_bdevs_operational": 1, 00:12:22.468 "base_bdevs_list": [ 00:12:22.468 { 00:12:22.468 "name": null, 00:12:22.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.468 "is_configured": false, 00:12:22.468 "data_offset": 0, 00:12:22.468 "data_size": 63488 00:12:22.469 }, 00:12:22.469 { 00:12:22.469 "name": "BaseBdev2", 00:12:22.469 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:22.469 "is_configured": true, 00:12:22.469 "data_offset": 2048, 00:12:22.469 "data_size": 63488 00:12:22.469 } 00:12:22.469 ] 00:12:22.469 }' 00:12:22.469 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:22.469 11:01:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.728 167.50 IOPS, 502.50 MiB/s [2024-10-29T11:01:28.226Z] 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.728 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.728 "name": "raid_bdev1", 00:12:22.728 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:22.728 "strip_size_kb": 0, 00:12:22.728 "state": "online", 00:12:22.728 "raid_level": "raid1", 00:12:22.728 "superblock": true, 00:12:22.728 "num_base_bdevs": 2, 00:12:22.728 "num_base_bdevs_discovered": 1, 00:12:22.728 "num_base_bdevs_operational": 1, 00:12:22.728 "base_bdevs_list": [ 00:12:22.728 { 00:12:22.728 "name": null, 00:12:22.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.728 "is_configured": false, 00:12:22.728 "data_offset": 0, 00:12:22.728 "data_size": 63488 00:12:22.728 }, 00:12:22.728 { 00:12:22.728 "name": "BaseBdev2", 00:12:22.728 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:22.728 "is_configured": true, 00:12:22.728 "data_offset": 2048, 00:12:22.728 "data_size": 63488 00:12:22.728 } 00:12:22.728 ] 00:12:22.728 }' 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:22.988 [2024-10-29 11:01:28.316637] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.988 11:01:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:22.988 [2024-10-29 11:01:28.363556] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:12:22.988 [2024-10-29 11:01:28.365449] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:22.988 [2024-10-29 11:01:28.482416] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:22.988 [2024-10-29 11:01:28.482937] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:23.247 [2024-10-29 11:01:28.701901] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:23.247 [2024-10-29 11:01:28.702166] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:23.815 [2024-10-29 11:01:29.043868] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:23.815 158.67 IOPS, 476.00 MiB/s [2024-10-29T11:01:29.313Z] [2024-10-29 11:01:29.272644] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:23.815 [2024-10-29 11:01:29.272813] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.075 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.075 "name": "raid_bdev1", 00:12:24.075 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:24.075 "strip_size_kb": 0, 00:12:24.075 "state": "online", 00:12:24.075 "raid_level": "raid1", 00:12:24.075 "superblock": true, 00:12:24.075 "num_base_bdevs": 2, 00:12:24.075 "num_base_bdevs_discovered": 2, 00:12:24.075 "num_base_bdevs_operational": 2, 00:12:24.075 "process": { 00:12:24.075 "type": "rebuild", 00:12:24.075 "target": "spare", 00:12:24.075 "progress": { 00:12:24.075 "blocks": 10240, 00:12:24.075 "percent": 16 00:12:24.075 } 00:12:24.075 }, 00:12:24.075 "base_bdevs_list": [ 00:12:24.075 { 00:12:24.075 "name": "spare", 00:12:24.076 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:24.076 "is_configured": true, 00:12:24.076 "data_offset": 2048, 00:12:24.076 "data_size": 63488 00:12:24.076 }, 00:12:24.076 { 00:12:24.076 "name": "BaseBdev2", 00:12:24.076 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:24.076 "is_configured": true, 00:12:24.076 "data_offset": 2048, 00:12:24.076 "data_size": 63488 00:12:24.076 } 00:12:24.076 ] 00:12:24.076 }' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:24.076 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=334 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.076 "name": "raid_bdev1", 00:12:24.076 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:24.076 "strip_size_kb": 0, 00:12:24.076 "state": "online", 00:12:24.076 "raid_level": "raid1", 00:12:24.076 "superblock": true, 00:12:24.076 "num_base_bdevs": 2, 00:12:24.076 "num_base_bdevs_discovered": 2, 00:12:24.076 "num_base_bdevs_operational": 2, 00:12:24.076 "process": { 00:12:24.076 "type": "rebuild", 00:12:24.076 "target": "spare", 00:12:24.076 "progress": { 00:12:24.076 "blocks": 10240, 00:12:24.076 "percent": 16 00:12:24.076 } 00:12:24.076 }, 00:12:24.076 "base_bdevs_list": [ 00:12:24.076 { 00:12:24.076 "name": "spare", 00:12:24.076 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:24.076 "is_configured": true, 00:12:24.076 "data_offset": 2048, 00:12:24.076 "data_size": 63488 00:12:24.076 }, 00:12:24.076 { 00:12:24.076 "name": "BaseBdev2", 00:12:24.076 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:24.076 "is_configured": true, 00:12:24.076 "data_offset": 2048, 00:12:24.076 "data_size": 63488 00:12:24.076 } 00:12:24.076 ] 00:12:24.076 }' 00:12:24.076 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.340 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.340 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.340 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.340 11:01:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:25.168 133.50 IOPS, 400.50 MiB/s [2024-10-29T11:01:30.666Z] [2024-10-29 11:01:30.391166] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:25.168 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:25.428 "name": "raid_bdev1", 00:12:25.428 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:25.428 "strip_size_kb": 0, 00:12:25.428 "state": "online", 00:12:25.428 "raid_level": "raid1", 00:12:25.428 "superblock": true, 00:12:25.428 "num_base_bdevs": 2, 00:12:25.428 "num_base_bdevs_discovered": 2, 00:12:25.428 "num_base_bdevs_operational": 2, 00:12:25.428 "process": { 00:12:25.428 "type": "rebuild", 00:12:25.428 "target": "spare", 00:12:25.428 "progress": { 00:12:25.428 "blocks": 30720, 00:12:25.428 "percent": 48 00:12:25.428 } 00:12:25.428 }, 00:12:25.428 "base_bdevs_list": [ 00:12:25.428 { 00:12:25.428 "name": "spare", 00:12:25.428 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:25.428 "is_configured": true, 00:12:25.428 "data_offset": 2048, 00:12:25.428 "data_size": 63488 00:12:25.428 }, 00:12:25.428 { 00:12:25.428 "name": "BaseBdev2", 00:12:25.428 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:25.428 "is_configured": true, 00:12:25.428 "data_offset": 2048, 00:12:25.428 "data_size": 63488 00:12:25.428 } 00:12:25.428 ] 00:12:25.428 }' 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:25.428 11:01:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:25.428 [2024-10-29 11:01:30.861454] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:25.948 118.60 IOPS, 355.80 MiB/s [2024-10-29T11:01:31.446Z] [2024-10-29 11:01:31.300986] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:25.948 [2024-10-29 11:01:31.301279] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:26.208 [2024-10-29 11:01:31.634368] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.468 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.468 "name": "raid_bdev1", 00:12:26.468 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:26.468 "strip_size_kb": 0, 00:12:26.468 "state": "online", 00:12:26.468 "raid_level": "raid1", 00:12:26.468 "superblock": true, 00:12:26.468 "num_base_bdevs": 2, 00:12:26.468 "num_base_bdevs_discovered": 2, 00:12:26.468 "num_base_bdevs_operational": 2, 00:12:26.468 "process": { 00:12:26.468 "type": "rebuild", 00:12:26.468 "target": "spare", 00:12:26.468 "progress": { 00:12:26.468 "blocks": 49152, 00:12:26.468 "percent": 77 00:12:26.468 } 00:12:26.468 }, 00:12:26.468 "base_bdevs_list": [ 00:12:26.468 { 00:12:26.468 "name": "spare", 00:12:26.468 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:26.468 "is_configured": true, 00:12:26.468 "data_offset": 2048, 00:12:26.468 "data_size": 63488 00:12:26.468 }, 00:12:26.468 { 00:12:26.468 "name": "BaseBdev2", 00:12:26.468 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:26.468 "is_configured": true, 00:12:26.468 "data_offset": 2048, 00:12:26.468 "data_size": 63488 00:12:26.468 } 00:12:26.469 ] 00:12:26.469 }' 00:12:26.469 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.469 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.469 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.469 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.469 11:01:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:26.988 107.00 IOPS, 321.00 MiB/s [2024-10-29T11:01:32.486Z] [2024-10-29 11:01:32.293561] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:12:26.988 [2024-10-29 11:01:32.293847] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:12:27.248 [2024-10-29 11:01:32.523340] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:27.248 [2024-10-29 11:01:32.628227] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:27.248 [2024-10-29 11:01:32.630260] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:27.509 "name": "raid_bdev1", 00:12:27.509 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:27.509 "strip_size_kb": 0, 00:12:27.509 "state": "online", 00:12:27.509 "raid_level": "raid1", 00:12:27.509 "superblock": true, 00:12:27.509 "num_base_bdevs": 2, 00:12:27.509 "num_base_bdevs_discovered": 2, 00:12:27.509 "num_base_bdevs_operational": 2, 00:12:27.509 "base_bdevs_list": [ 00:12:27.509 { 00:12:27.509 "name": "spare", 00:12:27.509 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:27.509 "is_configured": true, 00:12:27.509 "data_offset": 2048, 00:12:27.509 "data_size": 63488 00:12:27.509 }, 00:12:27.509 { 00:12:27.509 "name": "BaseBdev2", 00:12:27.509 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:27.509 "is_configured": true, 00:12:27.509 "data_offset": 2048, 00:12:27.509 "data_size": 63488 00:12:27.509 } 00:12:27.509 ] 00:12:27.509 }' 00:12:27.509 11:01:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:27.768 99.29 IOPS, 297.86 MiB/s [2024-10-29T11:01:33.266Z] 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.768 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:27.768 "name": "raid_bdev1", 00:12:27.768 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:27.768 "strip_size_kb": 0, 00:12:27.768 "state": "online", 00:12:27.768 "raid_level": "raid1", 00:12:27.768 "superblock": true, 00:12:27.768 "num_base_bdevs": 2, 00:12:27.768 "num_base_bdevs_discovered": 2, 00:12:27.768 "num_base_bdevs_operational": 2, 00:12:27.768 "base_bdevs_list": [ 00:12:27.768 { 00:12:27.768 "name": "spare", 00:12:27.768 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:27.768 "is_configured": true, 00:12:27.769 "data_offset": 2048, 00:12:27.769 "data_size": 63488 00:12:27.769 }, 00:12:27.769 { 00:12:27.769 "name": "BaseBdev2", 00:12:27.769 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:27.769 "is_configured": true, 00:12:27.769 "data_offset": 2048, 00:12:27.769 "data_size": 63488 00:12:27.769 } 00:12:27.769 ] 00:12:27.769 }' 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.769 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.029 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:28.029 "name": "raid_bdev1", 00:12:28.029 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:28.029 "strip_size_kb": 0, 00:12:28.029 "state": "online", 00:12:28.029 "raid_level": "raid1", 00:12:28.029 "superblock": true, 00:12:28.029 "num_base_bdevs": 2, 00:12:28.029 "num_base_bdevs_discovered": 2, 00:12:28.029 "num_base_bdevs_operational": 2, 00:12:28.029 "base_bdevs_list": [ 00:12:28.029 { 00:12:28.029 "name": "spare", 00:12:28.029 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:28.029 "is_configured": true, 00:12:28.029 "data_offset": 2048, 00:12:28.029 "data_size": 63488 00:12:28.029 }, 00:12:28.029 { 00:12:28.029 "name": "BaseBdev2", 00:12:28.029 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:28.029 "is_configured": true, 00:12:28.029 "data_offset": 2048, 00:12:28.029 "data_size": 63488 00:12:28.029 } 00:12:28.029 ] 00:12:28.029 }' 00:12:28.029 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:28.029 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.289 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:28.289 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.289 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.289 [2024-10-29 11:01:33.717701] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:28.289 [2024-10-29 11:01:33.717812] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:28.548 00:12:28.548 Latency(us) 00:12:28.548 [2024-10-29T11:01:34.046Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:28.548 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:28.548 raid_bdev1 : 7.77 95.22 285.66 0.00 0.00 14607.56 275.45 113099.68 00:12:28.548 [2024-10-29T11:01:34.046Z] =================================================================================================================== 00:12:28.548 [2024-10-29T11:01:34.046Z] Total : 95.22 285.66 0.00 0.00 14607.56 275.45 113099.68 00:12:28.548 [2024-10-29 11:01:33.812678] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:28.548 [2024-10-29 11:01:33.812747] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:28.548 [2024-10-29 11:01:33.812849] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:28.548 [2024-10-29 11:01:33.812936] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:28.548 { 00:12:28.548 "results": [ 00:12:28.548 { 00:12:28.548 "job": "raid_bdev1", 00:12:28.548 "core_mask": "0x1", 00:12:28.548 "workload": "randrw", 00:12:28.548 "percentage": 50, 00:12:28.548 "status": "finished", 00:12:28.548 "queue_depth": 2, 00:12:28.548 "io_size": 3145728, 00:12:28.548 "runtime": 7.771508, 00:12:28.548 "iops": 95.21961503481693, 00:12:28.548 "mibps": 285.6588451044508, 00:12:28.548 "io_failed": 0, 00:12:28.548 "io_timeout": 0, 00:12:28.548 "avg_latency_us": 14607.557597073055, 00:12:28.548 "min_latency_us": 275.45152838427947, 00:12:28.548 "max_latency_us": 113099.68209606987 00:12:28.548 } 00:12:28.548 ], 00:12:28.548 "core_count": 1 00:12:28.548 } 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:28.548 11:01:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:28.809 /dev/nbd0 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # local i 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # break 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:28.809 1+0 records in 00:12:28.809 1+0 records out 00:12:28.809 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000474769 s, 8.6 MB/s 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # size=4096 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # return 0 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:28.809 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:29.069 /dev/nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # local i 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # break 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:29.069 1+0 records in 00:12:29.069 1+0 records out 00:12:29.069 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000409546 s, 10.0 MB/s 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # size=4096 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # return 0 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:29.069 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:29.330 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.590 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.590 [2024-10-29 11:01:34.870415] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:29.591 [2024-10-29 11:01:34.870464] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:29.591 [2024-10-29 11:01:34.870502] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:12:29.591 [2024-10-29 11:01:34.870510] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:29.591 [2024-10-29 11:01:34.872644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:29.591 [2024-10-29 11:01:34.872679] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:29.591 [2024-10-29 11:01:34.872763] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:29.591 [2024-10-29 11:01:34.872811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:29.591 [2024-10-29 11:01:34.872932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:29.591 spare 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.591 [2024-10-29 11:01:34.972829] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:12:29.591 [2024-10-29 11:01:34.972863] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:29.591 [2024-10-29 11:01:34.973097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002af30 00:12:29.591 [2024-10-29 11:01:34.973224] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:12:29.591 [2024-10-29 11:01:34.973234] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:12:29.591 [2024-10-29 11:01:34.973381] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.591 11:01:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.591 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.591 "name": "raid_bdev1", 00:12:29.591 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:29.591 "strip_size_kb": 0, 00:12:29.591 "state": "online", 00:12:29.591 "raid_level": "raid1", 00:12:29.591 "superblock": true, 00:12:29.591 "num_base_bdevs": 2, 00:12:29.591 "num_base_bdevs_discovered": 2, 00:12:29.591 "num_base_bdevs_operational": 2, 00:12:29.591 "base_bdevs_list": [ 00:12:29.591 { 00:12:29.591 "name": "spare", 00:12:29.591 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:29.591 "is_configured": true, 00:12:29.591 "data_offset": 2048, 00:12:29.591 "data_size": 63488 00:12:29.591 }, 00:12:29.591 { 00:12:29.591 "name": "BaseBdev2", 00:12:29.591 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:29.591 "is_configured": true, 00:12:29.591 "data_offset": 2048, 00:12:29.591 "data_size": 63488 00:12:29.591 } 00:12:29.591 ] 00:12:29.591 }' 00:12:29.591 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.591 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.163 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.163 "name": "raid_bdev1", 00:12:30.163 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:30.163 "strip_size_kb": 0, 00:12:30.163 "state": "online", 00:12:30.163 "raid_level": "raid1", 00:12:30.163 "superblock": true, 00:12:30.163 "num_base_bdevs": 2, 00:12:30.163 "num_base_bdevs_discovered": 2, 00:12:30.163 "num_base_bdevs_operational": 2, 00:12:30.163 "base_bdevs_list": [ 00:12:30.163 { 00:12:30.163 "name": "spare", 00:12:30.163 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:30.163 "is_configured": true, 00:12:30.163 "data_offset": 2048, 00:12:30.163 "data_size": 63488 00:12:30.163 }, 00:12:30.163 { 00:12:30.163 "name": "BaseBdev2", 00:12:30.163 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:30.163 "is_configured": true, 00:12:30.163 "data_offset": 2048, 00:12:30.163 "data_size": 63488 00:12:30.163 } 00:12:30.164 ] 00:12:30.164 }' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.164 [2024-10-29 11:01:35.593264] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:30.164 "name": "raid_bdev1", 00:12:30.164 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:30.164 "strip_size_kb": 0, 00:12:30.164 "state": "online", 00:12:30.164 "raid_level": "raid1", 00:12:30.164 "superblock": true, 00:12:30.164 "num_base_bdevs": 2, 00:12:30.164 "num_base_bdevs_discovered": 1, 00:12:30.164 "num_base_bdevs_operational": 1, 00:12:30.164 "base_bdevs_list": [ 00:12:30.164 { 00:12:30.164 "name": null, 00:12:30.164 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:30.164 "is_configured": false, 00:12:30.164 "data_offset": 0, 00:12:30.164 "data_size": 63488 00:12:30.164 }, 00:12:30.164 { 00:12:30.164 "name": "BaseBdev2", 00:12:30.164 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:30.164 "is_configured": true, 00:12:30.164 "data_offset": 2048, 00:12:30.164 "data_size": 63488 00:12:30.164 } 00:12:30.164 ] 00:12:30.164 }' 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:30.164 11:01:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.782 11:01:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:30.782 11:01:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.782 11:01:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.782 [2024-10-29 11:01:36.016606] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.782 [2024-10-29 11:01:36.016824] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:30.782 [2024-10-29 11:01:36.016883] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:30.782 [2024-10-29 11:01:36.016944] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.782 [2024-10-29 11:01:36.022052] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b000 00:12:30.782 11:01:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.782 11:01:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:30.782 [2024-10-29 11:01:36.023977] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.722 "name": "raid_bdev1", 00:12:31.722 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:31.722 "strip_size_kb": 0, 00:12:31.722 "state": "online", 00:12:31.722 "raid_level": "raid1", 00:12:31.722 "superblock": true, 00:12:31.722 "num_base_bdevs": 2, 00:12:31.722 "num_base_bdevs_discovered": 2, 00:12:31.722 "num_base_bdevs_operational": 2, 00:12:31.722 "process": { 00:12:31.722 "type": "rebuild", 00:12:31.722 "target": "spare", 00:12:31.722 "progress": { 00:12:31.722 "blocks": 20480, 00:12:31.722 "percent": 32 00:12:31.722 } 00:12:31.722 }, 00:12:31.722 "base_bdevs_list": [ 00:12:31.722 { 00:12:31.722 "name": "spare", 00:12:31.722 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:31.722 "is_configured": true, 00:12:31.722 "data_offset": 2048, 00:12:31.722 "data_size": 63488 00:12:31.722 }, 00:12:31.722 { 00:12:31.722 "name": "BaseBdev2", 00:12:31.722 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:31.722 "is_configured": true, 00:12:31.722 "data_offset": 2048, 00:12:31.722 "data_size": 63488 00:12:31.722 } 00:12:31.722 ] 00:12:31.722 }' 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.722 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.722 [2024-10-29 11:01:37.184080] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.983 [2024-10-29 11:01:37.227812] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:31.983 [2024-10-29 11:01:37.227929] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:31.983 [2024-10-29 11:01:37.227986] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.983 [2024-10-29 11:01:37.228012] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.983 "name": "raid_bdev1", 00:12:31.983 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:31.983 "strip_size_kb": 0, 00:12:31.983 "state": "online", 00:12:31.983 "raid_level": "raid1", 00:12:31.983 "superblock": true, 00:12:31.983 "num_base_bdevs": 2, 00:12:31.983 "num_base_bdevs_discovered": 1, 00:12:31.983 "num_base_bdevs_operational": 1, 00:12:31.983 "base_bdevs_list": [ 00:12:31.983 { 00:12:31.983 "name": null, 00:12:31.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.983 "is_configured": false, 00:12:31.983 "data_offset": 0, 00:12:31.983 "data_size": 63488 00:12:31.983 }, 00:12:31.983 { 00:12:31.983 "name": "BaseBdev2", 00:12:31.983 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:31.983 "is_configured": true, 00:12:31.983 "data_offset": 2048, 00:12:31.983 "data_size": 63488 00:12:31.983 } 00:12:31.983 ] 00:12:31.983 }' 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.983 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.244 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:32.244 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:32.244 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.244 [2024-10-29 11:01:37.675930] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:32.244 [2024-10-29 11:01:37.675996] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:32.244 [2024-10-29 11:01:37.676017] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:32.244 [2024-10-29 11:01:37.676028] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:32.244 [2024-10-29 11:01:37.676463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:32.244 [2024-10-29 11:01:37.676516] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:32.244 [2024-10-29 11:01:37.676643] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:32.244 [2024-10-29 11:01:37.676659] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:32.244 [2024-10-29 11:01:37.676669] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:32.244 [2024-10-29 11:01:37.676697] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:32.244 spare 00:12:32.244 [2024-10-29 11:01:37.681740] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b0d0 00:12:32.244 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:32.244 11:01:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:32.244 [2024-10-29 11:01:37.683587] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.636 "name": "raid_bdev1", 00:12:33.636 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:33.636 "strip_size_kb": 0, 00:12:33.636 "state": "online", 00:12:33.636 "raid_level": "raid1", 00:12:33.636 "superblock": true, 00:12:33.636 "num_base_bdevs": 2, 00:12:33.636 "num_base_bdevs_discovered": 2, 00:12:33.636 "num_base_bdevs_operational": 2, 00:12:33.636 "process": { 00:12:33.636 "type": "rebuild", 00:12:33.636 "target": "spare", 00:12:33.636 "progress": { 00:12:33.636 "blocks": 20480, 00:12:33.636 "percent": 32 00:12:33.636 } 00:12:33.636 }, 00:12:33.636 "base_bdevs_list": [ 00:12:33.636 { 00:12:33.636 "name": "spare", 00:12:33.636 "uuid": "003c6d02-3bf4-5d18-a51d-81006e1d0a44", 00:12:33.636 "is_configured": true, 00:12:33.636 "data_offset": 2048, 00:12:33.636 "data_size": 63488 00:12:33.636 }, 00:12:33.636 { 00:12:33.636 "name": "BaseBdev2", 00:12:33.636 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:33.636 "is_configured": true, 00:12:33.636 "data_offset": 2048, 00:12:33.636 "data_size": 63488 00:12:33.636 } 00:12:33.636 ] 00:12:33.636 }' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.636 [2024-10-29 11:01:38.843784] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.636 [2024-10-29 11:01:38.887355] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:33.636 [2024-10-29 11:01:38.887477] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:33.636 [2024-10-29 11:01:38.887550] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:33.636 [2024-10-29 11:01:38.887579] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:33.636 "name": "raid_bdev1", 00:12:33.636 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:33.636 "strip_size_kb": 0, 00:12:33.636 "state": "online", 00:12:33.636 "raid_level": "raid1", 00:12:33.636 "superblock": true, 00:12:33.636 "num_base_bdevs": 2, 00:12:33.636 "num_base_bdevs_discovered": 1, 00:12:33.636 "num_base_bdevs_operational": 1, 00:12:33.636 "base_bdevs_list": [ 00:12:33.636 { 00:12:33.636 "name": null, 00:12:33.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.636 "is_configured": false, 00:12:33.636 "data_offset": 0, 00:12:33.636 "data_size": 63488 00:12:33.636 }, 00:12:33.636 { 00:12:33.636 "name": "BaseBdev2", 00:12:33.636 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:33.636 "is_configured": true, 00:12:33.636 "data_offset": 2048, 00:12:33.636 "data_size": 63488 00:12:33.636 } 00:12:33.636 ] 00:12:33.636 }' 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:33.636 11:01:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.896 "name": "raid_bdev1", 00:12:33.896 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:33.896 "strip_size_kb": 0, 00:12:33.896 "state": "online", 00:12:33.896 "raid_level": "raid1", 00:12:33.896 "superblock": true, 00:12:33.896 "num_base_bdevs": 2, 00:12:33.896 "num_base_bdevs_discovered": 1, 00:12:33.896 "num_base_bdevs_operational": 1, 00:12:33.896 "base_bdevs_list": [ 00:12:33.896 { 00:12:33.896 "name": null, 00:12:33.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.896 "is_configured": false, 00:12:33.896 "data_offset": 0, 00:12:33.896 "data_size": 63488 00:12:33.896 }, 00:12:33.896 { 00:12:33.896 "name": "BaseBdev2", 00:12:33.896 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:33.896 "is_configured": true, 00:12:33.896 "data_offset": 2048, 00:12:33.896 "data_size": 63488 00:12:33.896 } 00:12:33.896 ] 00:12:33.896 }' 00:12:33.896 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.156 [2024-10-29 11:01:39.467721] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:34.156 [2024-10-29 11:01:39.468019] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:34.156 [2024-10-29 11:01:39.468109] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:12:34.156 [2024-10-29 11:01:39.468159] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:34.156 [2024-10-29 11:01:39.468639] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:34.156 [2024-10-29 11:01:39.468724] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:34.156 [2024-10-29 11:01:39.468843] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:34.156 [2024-10-29 11:01:39.468870] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:34.156 [2024-10-29 11:01:39.468879] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:34.156 [2024-10-29 11:01:39.468899] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:34.156 BaseBdev1 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.156 11:01:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.094 "name": "raid_bdev1", 00:12:35.094 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:35.094 "strip_size_kb": 0, 00:12:35.094 "state": "online", 00:12:35.094 "raid_level": "raid1", 00:12:35.094 "superblock": true, 00:12:35.094 "num_base_bdevs": 2, 00:12:35.094 "num_base_bdevs_discovered": 1, 00:12:35.094 "num_base_bdevs_operational": 1, 00:12:35.094 "base_bdevs_list": [ 00:12:35.094 { 00:12:35.094 "name": null, 00:12:35.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.094 "is_configured": false, 00:12:35.094 "data_offset": 0, 00:12:35.094 "data_size": 63488 00:12:35.094 }, 00:12:35.094 { 00:12:35.094 "name": "BaseBdev2", 00:12:35.094 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:35.094 "is_configured": true, 00:12:35.094 "data_offset": 2048, 00:12:35.094 "data_size": 63488 00:12:35.094 } 00:12:35.094 ] 00:12:35.094 }' 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.094 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.664 "name": "raid_bdev1", 00:12:35.664 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:35.664 "strip_size_kb": 0, 00:12:35.664 "state": "online", 00:12:35.664 "raid_level": "raid1", 00:12:35.664 "superblock": true, 00:12:35.664 "num_base_bdevs": 2, 00:12:35.664 "num_base_bdevs_discovered": 1, 00:12:35.664 "num_base_bdevs_operational": 1, 00:12:35.664 "base_bdevs_list": [ 00:12:35.664 { 00:12:35.664 "name": null, 00:12:35.664 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.664 "is_configured": false, 00:12:35.664 "data_offset": 0, 00:12:35.664 "data_size": 63488 00:12:35.664 }, 00:12:35.664 { 00:12:35.664 "name": "BaseBdev2", 00:12:35.664 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:35.664 "is_configured": true, 00:12:35.664 "data_offset": 2048, 00:12:35.664 "data_size": 63488 00:12:35.664 } 00:12:35.664 ] 00:12:35.664 }' 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:35.664 11:01:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:35.664 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:35.664 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.665 [2024-10-29 11:01:41.045482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:35.665 [2024-10-29 11:01:41.045624] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:35.665 [2024-10-29 11:01:41.045639] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:35.665 request: 00:12:35.665 { 00:12:35.665 "base_bdev": "BaseBdev1", 00:12:35.665 "raid_bdev": "raid_bdev1", 00:12:35.665 "method": "bdev_raid_add_base_bdev", 00:12:35.665 "req_id": 1 00:12:35.665 } 00:12:35.665 Got JSON-RPC error response 00:12:35.665 response: 00:12:35.665 { 00:12:35.665 "code": -22, 00:12:35.665 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:35.665 } 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:35.665 11:01:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.609 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.873 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.873 "name": "raid_bdev1", 00:12:36.873 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:36.873 "strip_size_kb": 0, 00:12:36.873 "state": "online", 00:12:36.873 "raid_level": "raid1", 00:12:36.873 "superblock": true, 00:12:36.873 "num_base_bdevs": 2, 00:12:36.873 "num_base_bdevs_discovered": 1, 00:12:36.873 "num_base_bdevs_operational": 1, 00:12:36.873 "base_bdevs_list": [ 00:12:36.873 { 00:12:36.873 "name": null, 00:12:36.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.873 "is_configured": false, 00:12:36.873 "data_offset": 0, 00:12:36.873 "data_size": 63488 00:12:36.873 }, 00:12:36.873 { 00:12:36.873 "name": "BaseBdev2", 00:12:36.873 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:36.873 "is_configured": true, 00:12:36.873 "data_offset": 2048, 00:12:36.873 "data_size": 63488 00:12:36.873 } 00:12:36.873 ] 00:12:36.873 }' 00:12:36.873 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.873 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:37.134 "name": "raid_bdev1", 00:12:37.134 "uuid": "ccbe68ce-1cce-46ca-b46d-5b771318b6a6", 00:12:37.134 "strip_size_kb": 0, 00:12:37.134 "state": "online", 00:12:37.134 "raid_level": "raid1", 00:12:37.134 "superblock": true, 00:12:37.134 "num_base_bdevs": 2, 00:12:37.134 "num_base_bdevs_discovered": 1, 00:12:37.134 "num_base_bdevs_operational": 1, 00:12:37.134 "base_bdevs_list": [ 00:12:37.134 { 00:12:37.134 "name": null, 00:12:37.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.134 "is_configured": false, 00:12:37.134 "data_offset": 0, 00:12:37.134 "data_size": 63488 00:12:37.134 }, 00:12:37.134 { 00:12:37.134 "name": "BaseBdev2", 00:12:37.134 "uuid": "26002a7a-46de-5ab8-a075-8ea027b98f98", 00:12:37.134 "is_configured": true, 00:12:37.134 "data_offset": 2048, 00:12:37.134 "data_size": 63488 00:12:37.134 } 00:12:37.134 ] 00:12:37.134 }' 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:37.134 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87643 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@952 -- # '[' -z 87643 ']' 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # kill -0 87643 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@957 -- # uname 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:12:37.394 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 87643 00:12:37.394 killing process with pid 87643 00:12:37.394 Received shutdown signal, test time was about 16.675978 seconds 00:12:37.394 00:12:37.394 Latency(us) 00:12:37.395 [2024-10-29T11:01:42.893Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:37.395 [2024-10-29T11:01:42.893Z] =================================================================================================================== 00:12:37.395 [2024-10-29T11:01:42.893Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:37.395 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:12:37.395 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:12:37.395 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@970 -- # echo 'killing process with pid 87643' 00:12:37.395 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@971 -- # kill 87643 00:12:37.395 [2024-10-29 11:01:42.696814] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:37.395 [2024-10-29 11:01:42.696946] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:37.395 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@976 -- # wait 87643 00:12:37.395 [2024-10-29 11:01:42.696997] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:37.395 [2024-10-29 11:01:42.697008] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:12:37.395 [2024-10-29 11:01:42.722685] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:37.655 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:37.655 ************************************ 00:12:37.655 END TEST raid_rebuild_test_sb_io 00:12:37.655 ************************************ 00:12:37.656 00:12:37.656 real 0m18.555s 00:12:37.656 user 0m24.709s 00:12:37.656 sys 0m2.095s 00:12:37.656 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1128 -- # xtrace_disable 00:12:37.656 11:01:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.656 11:01:42 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:37.656 11:01:42 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:12:37.656 11:01:42 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:12:37.656 11:01:42 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:12:37.656 11:01:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:37.656 ************************************ 00:12:37.656 START TEST raid_rebuild_test 00:12:37.656 ************************************ 00:12:37.656 11:01:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 4 false false true 00:12:37.656 11:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:37.656 11:01:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=88316 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 88316 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@833 -- # '[' -z 88316 ']' 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:37.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:12:37.656 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:37.656 [2024-10-29 11:01:43.095662] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:12:37.656 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:37.656 Zero copy mechanism will not be used. 00:12:37.656 [2024-10-29 11:01:43.095868] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88316 ] 00:12:37.917 [2024-10-29 11:01:43.265344] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:37.917 [2024-10-29 11:01:43.289195] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:37.917 [2024-10-29 11:01:43.329962] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:37.917 [2024-10-29 11:01:43.329997] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@866 -- # return 0 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.488 BaseBdev1_malloc 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.488 [2024-10-29 11:01:43.935400] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:38.488 [2024-10-29 11:01:43.935462] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.488 [2024-10-29 11:01:43.935488] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:38.488 [2024-10-29 11:01:43.935508] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.488 [2024-10-29 11:01:43.937583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.488 [2024-10-29 11:01:43.937630] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:38.488 BaseBdev1 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.488 BaseBdev2_malloc 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.488 [2024-10-29 11:01:43.963795] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:38.488 [2024-10-29 11:01:43.963841] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.488 [2024-10-29 11:01:43.963861] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:38.488 [2024-10-29 11:01:43.963870] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.488 [2024-10-29 11:01:43.965890] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.488 [2024-10-29 11:01:43.965965] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:38.488 BaseBdev2 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.488 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.750 BaseBdev3_malloc 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.750 [2024-10-29 11:01:43.992282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:38.750 [2024-10-29 11:01:43.992328] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.750 [2024-10-29 11:01:43.992351] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:38.750 [2024-10-29 11:01:43.992375] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.750 [2024-10-29 11:01:43.994392] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.750 [2024-10-29 11:01:43.994419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:38.750 BaseBdev3 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.750 11:01:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.750 BaseBdev4_malloc 00:12:38.750 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.750 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:38.750 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.750 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.750 [2024-10-29 11:01:44.037287] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:38.750 [2024-10-29 11:01:44.037400] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.750 [2024-10-29 11:01:44.037445] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:38.751 [2024-10-29 11:01:44.037464] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.751 [2024-10-29 11:01:44.041665] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.751 [2024-10-29 11:01:44.041727] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:38.751 BaseBdev4 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.751 spare_malloc 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.751 spare_delay 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.751 [2024-10-29 11:01:44.079168] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:38.751 [2024-10-29 11:01:44.079213] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:38.751 [2024-10-29 11:01:44.079250] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:38.751 [2024-10-29 11:01:44.079258] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:38.751 [2024-10-29 11:01:44.081307] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:38.751 [2024-10-29 11:01:44.081339] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:38.751 spare 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.751 [2024-10-29 11:01:44.091209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:38.751 [2024-10-29 11:01:44.093044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:38.751 [2024-10-29 11:01:44.093170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:38.751 [2024-10-29 11:01:44.093217] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:38.751 [2024-10-29 11:01:44.093313] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:38.751 [2024-10-29 11:01:44.093323] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:38.751 [2024-10-29 11:01:44.093593] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:38.751 [2024-10-29 11:01:44.093726] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:38.751 [2024-10-29 11:01:44.093739] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:38.751 [2024-10-29 11:01:44.093855] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:38.751 "name": "raid_bdev1", 00:12:38.751 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:38.751 "strip_size_kb": 0, 00:12:38.751 "state": "online", 00:12:38.751 "raid_level": "raid1", 00:12:38.751 "superblock": false, 00:12:38.751 "num_base_bdevs": 4, 00:12:38.751 "num_base_bdevs_discovered": 4, 00:12:38.751 "num_base_bdevs_operational": 4, 00:12:38.751 "base_bdevs_list": [ 00:12:38.751 { 00:12:38.751 "name": "BaseBdev1", 00:12:38.751 "uuid": "a3daa702-1ae4-58ed-9551-2312e923d880", 00:12:38.751 "is_configured": true, 00:12:38.751 "data_offset": 0, 00:12:38.751 "data_size": 65536 00:12:38.751 }, 00:12:38.751 { 00:12:38.751 "name": "BaseBdev2", 00:12:38.751 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:38.751 "is_configured": true, 00:12:38.751 "data_offset": 0, 00:12:38.751 "data_size": 65536 00:12:38.751 }, 00:12:38.751 { 00:12:38.751 "name": "BaseBdev3", 00:12:38.751 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:38.751 "is_configured": true, 00:12:38.751 "data_offset": 0, 00:12:38.751 "data_size": 65536 00:12:38.751 }, 00:12:38.751 { 00:12:38.751 "name": "BaseBdev4", 00:12:38.751 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:38.751 "is_configured": true, 00:12:38.751 "data_offset": 0, 00:12:38.751 "data_size": 65536 00:12:38.751 } 00:12:38.751 ] 00:12:38.751 }' 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:38.751 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.012 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:39.012 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:39.012 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.012 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.273 [2024-10-29 11:01:44.514758] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:39.273 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:39.534 [2024-10-29 11:01:44.774131] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:39.534 /dev/nbd0 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:39.534 1+0 records in 00:12:39.534 1+0 records out 00:12:39.534 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000447874 s, 9.1 MB/s 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:39.534 11:01:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:44.815 65536+0 records in 00:12:44.815 65536+0 records out 00:12:44.815 33554432 bytes (34 MB, 32 MiB) copied, 4.65731 s, 7.2 MB/s 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:44.815 [2024-10-29 11:01:49.733402] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:44.815 [2024-10-29 11:01:49.752225] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.815 "name": "raid_bdev1", 00:12:44.815 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:44.815 "strip_size_kb": 0, 00:12:44.815 "state": "online", 00:12:44.815 "raid_level": "raid1", 00:12:44.815 "superblock": false, 00:12:44.815 "num_base_bdevs": 4, 00:12:44.815 "num_base_bdevs_discovered": 3, 00:12:44.815 "num_base_bdevs_operational": 3, 00:12:44.815 "base_bdevs_list": [ 00:12:44.815 { 00:12:44.815 "name": null, 00:12:44.815 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.815 "is_configured": false, 00:12:44.815 "data_offset": 0, 00:12:44.815 "data_size": 65536 00:12:44.815 }, 00:12:44.815 { 00:12:44.815 "name": "BaseBdev2", 00:12:44.815 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:44.815 "is_configured": true, 00:12:44.815 "data_offset": 0, 00:12:44.815 "data_size": 65536 00:12:44.815 }, 00:12:44.815 { 00:12:44.815 "name": "BaseBdev3", 00:12:44.815 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:44.815 "is_configured": true, 00:12:44.815 "data_offset": 0, 00:12:44.815 "data_size": 65536 00:12:44.815 }, 00:12:44.815 { 00:12:44.815 "name": "BaseBdev4", 00:12:44.815 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:44.815 "is_configured": true, 00:12:44.815 "data_offset": 0, 00:12:44.815 "data_size": 65536 00:12:44.815 } 00:12:44.815 ] 00:12:44.815 }' 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.815 11:01:49 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:44.815 11:01:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:44.815 11:01:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.815 11:01:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:44.815 [2024-10-29 11:01:50.155623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:44.815 [2024-10-29 11:01:50.159692] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09bd0 00:12:44.815 11:01:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.815 11:01:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:44.815 [2024-10-29 11:01:50.161583] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.755 "name": "raid_bdev1", 00:12:45.755 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:45.755 "strip_size_kb": 0, 00:12:45.755 "state": "online", 00:12:45.755 "raid_level": "raid1", 00:12:45.755 "superblock": false, 00:12:45.755 "num_base_bdevs": 4, 00:12:45.755 "num_base_bdevs_discovered": 4, 00:12:45.755 "num_base_bdevs_operational": 4, 00:12:45.755 "process": { 00:12:45.755 "type": "rebuild", 00:12:45.755 "target": "spare", 00:12:45.755 "progress": { 00:12:45.755 "blocks": 20480, 00:12:45.755 "percent": 31 00:12:45.755 } 00:12:45.755 }, 00:12:45.755 "base_bdevs_list": [ 00:12:45.755 { 00:12:45.755 "name": "spare", 00:12:45.755 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:45.755 "is_configured": true, 00:12:45.755 "data_offset": 0, 00:12:45.755 "data_size": 65536 00:12:45.755 }, 00:12:45.755 { 00:12:45.755 "name": "BaseBdev2", 00:12:45.755 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:45.755 "is_configured": true, 00:12:45.755 "data_offset": 0, 00:12:45.755 "data_size": 65536 00:12:45.755 }, 00:12:45.755 { 00:12:45.755 "name": "BaseBdev3", 00:12:45.755 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:45.755 "is_configured": true, 00:12:45.755 "data_offset": 0, 00:12:45.755 "data_size": 65536 00:12:45.755 }, 00:12:45.755 { 00:12:45.755 "name": "BaseBdev4", 00:12:45.755 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:45.755 "is_configured": true, 00:12:45.755 "data_offset": 0, 00:12:45.755 "data_size": 65536 00:12:45.755 } 00:12:45.755 ] 00:12:45.755 }' 00:12:45.755 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.015 [2024-10-29 11:01:51.319718] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:46.015 [2024-10-29 11:01:51.365852] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:46.015 [2024-10-29 11:01:51.365947] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:46.015 [2024-10-29 11:01:51.366000] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:46.015 [2024-10-29 11:01:51.366020] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.015 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.016 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.016 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.016 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:46.016 "name": "raid_bdev1", 00:12:46.016 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:46.016 "strip_size_kb": 0, 00:12:46.016 "state": "online", 00:12:46.016 "raid_level": "raid1", 00:12:46.016 "superblock": false, 00:12:46.016 "num_base_bdevs": 4, 00:12:46.016 "num_base_bdevs_discovered": 3, 00:12:46.016 "num_base_bdevs_operational": 3, 00:12:46.016 "base_bdevs_list": [ 00:12:46.016 { 00:12:46.016 "name": null, 00:12:46.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.016 "is_configured": false, 00:12:46.016 "data_offset": 0, 00:12:46.016 "data_size": 65536 00:12:46.016 }, 00:12:46.016 { 00:12:46.016 "name": "BaseBdev2", 00:12:46.016 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:46.016 "is_configured": true, 00:12:46.016 "data_offset": 0, 00:12:46.016 "data_size": 65536 00:12:46.016 }, 00:12:46.016 { 00:12:46.016 "name": "BaseBdev3", 00:12:46.016 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:46.016 "is_configured": true, 00:12:46.016 "data_offset": 0, 00:12:46.016 "data_size": 65536 00:12:46.016 }, 00:12:46.016 { 00:12:46.016 "name": "BaseBdev4", 00:12:46.016 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:46.016 "is_configured": true, 00:12:46.016 "data_offset": 0, 00:12:46.016 "data_size": 65536 00:12:46.016 } 00:12:46.016 ] 00:12:46.016 }' 00:12:46.016 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:46.016 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.586 "name": "raid_bdev1", 00:12:46.586 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:46.586 "strip_size_kb": 0, 00:12:46.586 "state": "online", 00:12:46.586 "raid_level": "raid1", 00:12:46.586 "superblock": false, 00:12:46.586 "num_base_bdevs": 4, 00:12:46.586 "num_base_bdevs_discovered": 3, 00:12:46.586 "num_base_bdevs_operational": 3, 00:12:46.586 "base_bdevs_list": [ 00:12:46.586 { 00:12:46.586 "name": null, 00:12:46.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.586 "is_configured": false, 00:12:46.586 "data_offset": 0, 00:12:46.586 "data_size": 65536 00:12:46.586 }, 00:12:46.586 { 00:12:46.586 "name": "BaseBdev2", 00:12:46.586 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:46.586 "is_configured": true, 00:12:46.586 "data_offset": 0, 00:12:46.586 "data_size": 65536 00:12:46.586 }, 00:12:46.586 { 00:12:46.586 "name": "BaseBdev3", 00:12:46.586 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:46.586 "is_configured": true, 00:12:46.586 "data_offset": 0, 00:12:46.586 "data_size": 65536 00:12:46.586 }, 00:12:46.586 { 00:12:46.586 "name": "BaseBdev4", 00:12:46.586 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:46.586 "is_configured": true, 00:12:46.586 "data_offset": 0, 00:12:46.586 "data_size": 65536 00:12:46.586 } 00:12:46.586 ] 00:12:46.586 }' 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.586 [2024-10-29 11:01:51.953318] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:46.586 [2024-10-29 11:01:51.957302] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.586 11:01:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:46.586 [2024-10-29 11:01:51.959134] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.525 11:01:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.525 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.525 "name": "raid_bdev1", 00:12:47.525 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:47.525 "strip_size_kb": 0, 00:12:47.525 "state": "online", 00:12:47.525 "raid_level": "raid1", 00:12:47.525 "superblock": false, 00:12:47.525 "num_base_bdevs": 4, 00:12:47.525 "num_base_bdevs_discovered": 4, 00:12:47.525 "num_base_bdevs_operational": 4, 00:12:47.525 "process": { 00:12:47.525 "type": "rebuild", 00:12:47.525 "target": "spare", 00:12:47.525 "progress": { 00:12:47.525 "blocks": 20480, 00:12:47.525 "percent": 31 00:12:47.525 } 00:12:47.525 }, 00:12:47.525 "base_bdevs_list": [ 00:12:47.525 { 00:12:47.525 "name": "spare", 00:12:47.525 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:47.525 "is_configured": true, 00:12:47.525 "data_offset": 0, 00:12:47.525 "data_size": 65536 00:12:47.525 }, 00:12:47.525 { 00:12:47.525 "name": "BaseBdev2", 00:12:47.525 "uuid": "50c637f5-c83d-5e5e-86e2-f642d5dd0776", 00:12:47.525 "is_configured": true, 00:12:47.525 "data_offset": 0, 00:12:47.525 "data_size": 65536 00:12:47.525 }, 00:12:47.525 { 00:12:47.525 "name": "BaseBdev3", 00:12:47.525 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:47.525 "is_configured": true, 00:12:47.526 "data_offset": 0, 00:12:47.526 "data_size": 65536 00:12:47.526 }, 00:12:47.526 { 00:12:47.526 "name": "BaseBdev4", 00:12:47.526 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:47.526 "is_configured": true, 00:12:47.526 "data_offset": 0, 00:12:47.526 "data_size": 65536 00:12:47.526 } 00:12:47.526 ] 00:12:47.526 }' 00:12:47.526 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.785 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.785 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.786 [2024-10-29 11:01:53.119879] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:47.786 [2024-10-29 11:01:53.162903] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09ca0 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.786 "name": "raid_bdev1", 00:12:47.786 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:47.786 "strip_size_kb": 0, 00:12:47.786 "state": "online", 00:12:47.786 "raid_level": "raid1", 00:12:47.786 "superblock": false, 00:12:47.786 "num_base_bdevs": 4, 00:12:47.786 "num_base_bdevs_discovered": 3, 00:12:47.786 "num_base_bdevs_operational": 3, 00:12:47.786 "process": { 00:12:47.786 "type": "rebuild", 00:12:47.786 "target": "spare", 00:12:47.786 "progress": { 00:12:47.786 "blocks": 24576, 00:12:47.786 "percent": 37 00:12:47.786 } 00:12:47.786 }, 00:12:47.786 "base_bdevs_list": [ 00:12:47.786 { 00:12:47.786 "name": "spare", 00:12:47.786 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:47.786 "is_configured": true, 00:12:47.786 "data_offset": 0, 00:12:47.786 "data_size": 65536 00:12:47.786 }, 00:12:47.786 { 00:12:47.786 "name": null, 00:12:47.786 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.786 "is_configured": false, 00:12:47.786 "data_offset": 0, 00:12:47.786 "data_size": 65536 00:12:47.786 }, 00:12:47.786 { 00:12:47.786 "name": "BaseBdev3", 00:12:47.786 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:47.786 "is_configured": true, 00:12:47.786 "data_offset": 0, 00:12:47.786 "data_size": 65536 00:12:47.786 }, 00:12:47.786 { 00:12:47.786 "name": "BaseBdev4", 00:12:47.786 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:47.786 "is_configured": true, 00:12:47.786 "data_offset": 0, 00:12:47.786 "data_size": 65536 00:12:47.786 } 00:12:47.786 ] 00:12:47.786 }' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=358 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.786 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.046 "name": "raid_bdev1", 00:12:48.046 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:48.046 "strip_size_kb": 0, 00:12:48.046 "state": "online", 00:12:48.046 "raid_level": "raid1", 00:12:48.046 "superblock": false, 00:12:48.046 "num_base_bdevs": 4, 00:12:48.046 "num_base_bdevs_discovered": 3, 00:12:48.046 "num_base_bdevs_operational": 3, 00:12:48.046 "process": { 00:12:48.046 "type": "rebuild", 00:12:48.046 "target": "spare", 00:12:48.046 "progress": { 00:12:48.046 "blocks": 26624, 00:12:48.046 "percent": 40 00:12:48.046 } 00:12:48.046 }, 00:12:48.046 "base_bdevs_list": [ 00:12:48.046 { 00:12:48.046 "name": "spare", 00:12:48.046 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:48.046 "is_configured": true, 00:12:48.046 "data_offset": 0, 00:12:48.046 "data_size": 65536 00:12:48.046 }, 00:12:48.046 { 00:12:48.046 "name": null, 00:12:48.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.046 "is_configured": false, 00:12:48.046 "data_offset": 0, 00:12:48.046 "data_size": 65536 00:12:48.046 }, 00:12:48.046 { 00:12:48.046 "name": "BaseBdev3", 00:12:48.046 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:48.046 "is_configured": true, 00:12:48.046 "data_offset": 0, 00:12:48.046 "data_size": 65536 00:12:48.046 }, 00:12:48.046 { 00:12:48.046 "name": "BaseBdev4", 00:12:48.046 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:48.046 "is_configured": true, 00:12:48.046 "data_offset": 0, 00:12:48.046 "data_size": 65536 00:12:48.046 } 00:12:48.046 ] 00:12:48.046 }' 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:48.046 11:01:53 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.987 "name": "raid_bdev1", 00:12:48.987 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:48.987 "strip_size_kb": 0, 00:12:48.987 "state": "online", 00:12:48.987 "raid_level": "raid1", 00:12:48.987 "superblock": false, 00:12:48.987 "num_base_bdevs": 4, 00:12:48.987 "num_base_bdevs_discovered": 3, 00:12:48.987 "num_base_bdevs_operational": 3, 00:12:48.987 "process": { 00:12:48.987 "type": "rebuild", 00:12:48.987 "target": "spare", 00:12:48.987 "progress": { 00:12:48.987 "blocks": 49152, 00:12:48.987 "percent": 75 00:12:48.987 } 00:12:48.987 }, 00:12:48.987 "base_bdevs_list": [ 00:12:48.987 { 00:12:48.987 "name": "spare", 00:12:48.987 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:48.987 "is_configured": true, 00:12:48.987 "data_offset": 0, 00:12:48.987 "data_size": 65536 00:12:48.987 }, 00:12:48.987 { 00:12:48.987 "name": null, 00:12:48.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.987 "is_configured": false, 00:12:48.987 "data_offset": 0, 00:12:48.987 "data_size": 65536 00:12:48.987 }, 00:12:48.987 { 00:12:48.987 "name": "BaseBdev3", 00:12:48.987 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:48.987 "is_configured": true, 00:12:48.987 "data_offset": 0, 00:12:48.987 "data_size": 65536 00:12:48.987 }, 00:12:48.987 { 00:12:48.987 "name": "BaseBdev4", 00:12:48.987 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:48.987 "is_configured": true, 00:12:48.987 "data_offset": 0, 00:12:48.987 "data_size": 65536 00:12:48.987 } 00:12:48.987 ] 00:12:48.987 }' 00:12:48.987 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.245 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.245 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.245 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.245 11:01:54 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:49.814 [2024-10-29 11:01:55.169464] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:49.814 [2024-10-29 11:01:55.169528] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:49.814 [2024-10-29 11:01:55.169567] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.384 "name": "raid_bdev1", 00:12:50.384 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:50.384 "strip_size_kb": 0, 00:12:50.384 "state": "online", 00:12:50.384 "raid_level": "raid1", 00:12:50.384 "superblock": false, 00:12:50.384 "num_base_bdevs": 4, 00:12:50.384 "num_base_bdevs_discovered": 3, 00:12:50.384 "num_base_bdevs_operational": 3, 00:12:50.384 "base_bdevs_list": [ 00:12:50.384 { 00:12:50.384 "name": "spare", 00:12:50.384 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": null, 00:12:50.384 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.384 "is_configured": false, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": "BaseBdev3", 00:12:50.384 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": "BaseBdev4", 00:12:50.384 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 } 00:12:50.384 ] 00:12:50.384 }' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.384 "name": "raid_bdev1", 00:12:50.384 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:50.384 "strip_size_kb": 0, 00:12:50.384 "state": "online", 00:12:50.384 "raid_level": "raid1", 00:12:50.384 "superblock": false, 00:12:50.384 "num_base_bdevs": 4, 00:12:50.384 "num_base_bdevs_discovered": 3, 00:12:50.384 "num_base_bdevs_operational": 3, 00:12:50.384 "base_bdevs_list": [ 00:12:50.384 { 00:12:50.384 "name": "spare", 00:12:50.384 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": null, 00:12:50.384 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.384 "is_configured": false, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": "BaseBdev3", 00:12:50.384 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 }, 00:12:50.384 { 00:12:50.384 "name": "BaseBdev4", 00:12:50.384 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:50.384 "is_configured": true, 00:12:50.384 "data_offset": 0, 00:12:50.384 "data_size": 65536 00:12:50.384 } 00:12:50.384 ] 00:12:50.384 }' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:50.384 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.385 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.644 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:50.644 "name": "raid_bdev1", 00:12:50.644 "uuid": "1d40b2b5-1e14-4c02-899f-7f2039704b12", 00:12:50.644 "strip_size_kb": 0, 00:12:50.644 "state": "online", 00:12:50.644 "raid_level": "raid1", 00:12:50.644 "superblock": false, 00:12:50.644 "num_base_bdevs": 4, 00:12:50.644 "num_base_bdevs_discovered": 3, 00:12:50.644 "num_base_bdevs_operational": 3, 00:12:50.644 "base_bdevs_list": [ 00:12:50.644 { 00:12:50.644 "name": "spare", 00:12:50.644 "uuid": "96ef5869-a328-5090-a296-70a5c7f5cbab", 00:12:50.644 "is_configured": true, 00:12:50.644 "data_offset": 0, 00:12:50.644 "data_size": 65536 00:12:50.644 }, 00:12:50.644 { 00:12:50.644 "name": null, 00:12:50.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.644 "is_configured": false, 00:12:50.644 "data_offset": 0, 00:12:50.644 "data_size": 65536 00:12:50.644 }, 00:12:50.644 { 00:12:50.644 "name": "BaseBdev3", 00:12:50.644 "uuid": "0a1a7a60-dbfc-53d1-9dd2-707ff8a79b71", 00:12:50.644 "is_configured": true, 00:12:50.644 "data_offset": 0, 00:12:50.644 "data_size": 65536 00:12:50.644 }, 00:12:50.644 { 00:12:50.644 "name": "BaseBdev4", 00:12:50.644 "uuid": "c9396c52-e034-5923-985b-324339b80ea6", 00:12:50.644 "is_configured": true, 00:12:50.644 "data_offset": 0, 00:12:50.644 "data_size": 65536 00:12:50.644 } 00:12:50.644 ] 00:12:50.644 }' 00:12:50.644 11:01:55 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:50.644 11:01:55 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.904 [2024-10-29 11:01:56.239831] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:50.904 [2024-10-29 11:01:56.239900] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:50.904 [2024-10-29 11:01:56.240038] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:50.904 [2024-10-29 11:01:56.240152] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:50.904 [2024-10-29 11:01:56.240201] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:50.904 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:51.164 /dev/nbd0 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:51.164 1+0 records in 00:12:51.164 1+0 records out 00:12:51.164 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000545454 s, 7.5 MB/s 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:12:51.164 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:51.165 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:51.165 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:12:51.165 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:51.165 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:51.165 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:51.425 /dev/nbd1 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # break 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:51.425 1+0 records in 00:12:51.425 1+0 records out 00:12:51.425 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000440839 s, 9.3 MB/s 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:51.425 11:01:56 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:51.685 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:51.944 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:51.944 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 88316 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@952 -- # '[' -z 88316 ']' 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # kill -0 88316 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@957 -- # uname 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 88316 00:12:51.945 killing process with pid 88316 00:12:51.945 Received shutdown signal, test time was about 60.000000 seconds 00:12:51.945 00:12:51.945 Latency(us) 00:12:51.945 [2024-10-29T11:01:57.443Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:51.945 [2024-10-29T11:01:57.443Z] =================================================================================================================== 00:12:51.945 [2024-10-29T11:01:57.443Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 88316' 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@971 -- # kill 88316 00:12:51.945 [2024-10-29 11:01:57.320665] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:51.945 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@976 -- # wait 88316 00:12:51.945 [2024-10-29 11:01:57.369519] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:52.205 00:12:52.205 real 0m14.573s 00:12:52.205 user 0m16.900s 00:12:52.205 sys 0m2.730s 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:12:52.205 ************************************ 00:12:52.205 END TEST raid_rebuild_test 00:12:52.205 ************************************ 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.205 11:01:57 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:52.205 11:01:57 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:12:52.205 11:01:57 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:12:52.205 11:01:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:52.205 ************************************ 00:12:52.205 START TEST raid_rebuild_test_sb 00:12:52.205 ************************************ 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 4 true false true 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88737 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88737 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@833 -- # '[' -z 88737 ']' 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:52.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:12:52.205 11:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:52.465 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:52.465 Zero copy mechanism will not be used. 00:12:52.465 [2024-10-29 11:01:57.742716] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:12:52.465 [2024-10-29 11:01:57.742933] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88737 ] 00:12:52.465 [2024-10-29 11:01:57.911048] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:52.465 [2024-10-29 11:01:57.935045] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:52.725 [2024-10-29 11:01:57.976863] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:52.725 [2024-10-29 11:01:57.976940] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@866 -- # return 0 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.295 BaseBdev1_malloc 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.295 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.566639] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:53.296 [2024-10-29 11:01:58.566741] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.296 [2024-10-29 11:01:58.566772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:53.296 [2024-10-29 11:01:58.566793] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.296 [2024-10-29 11:01:58.568941] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.296 [2024-10-29 11:01:58.568982] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:53.296 BaseBdev1 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 BaseBdev2_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.594809] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:53.296 [2024-10-29 11:01:58.594855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.296 [2024-10-29 11:01:58.594891] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:53.296 [2024-10-29 11:01:58.594899] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.296 [2024-10-29 11:01:58.596885] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.296 [2024-10-29 11:01:58.596965] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:53.296 BaseBdev2 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 BaseBdev3_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.623005] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:53.296 [2024-10-29 11:01:58.623050] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.296 [2024-10-29 11:01:58.623089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:53.296 [2024-10-29 11:01:58.623097] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.296 [2024-10-29 11:01:58.625084] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.296 [2024-10-29 11:01:58.625117] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:53.296 BaseBdev3 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 BaseBdev4_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.668183] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:53.296 [2024-10-29 11:01:58.668282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.296 [2024-10-29 11:01:58.668331] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:53.296 [2024-10-29 11:01:58.668353] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.296 [2024-10-29 11:01:58.672483] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.296 [2024-10-29 11:01:58.672532] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:53.296 BaseBdev4 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 spare_malloc 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 spare_delay 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.709661] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:53.296 [2024-10-29 11:01:58.709708] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:53.296 [2024-10-29 11:01:58.709744] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:53.296 [2024-10-29 11:01:58.709752] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:53.296 [2024-10-29 11:01:58.711790] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:53.296 [2024-10-29 11:01:58.711874] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:53.296 spare 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 [2024-10-29 11:01:58.721703] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:53.296 [2024-10-29 11:01:58.723454] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:53.296 [2024-10-29 11:01:58.723518] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:53.296 [2024-10-29 11:01:58.723560] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:53.296 [2024-10-29 11:01:58.723732] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:12:53.296 [2024-10-29 11:01:58.723748] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:53.296 [2024-10-29 11:01:58.723978] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:12:53.296 [2024-10-29 11:01:58.724116] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:12:53.296 [2024-10-29 11:01:58.724140] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:12:53.296 [2024-10-29 11:01:58.724251] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.296 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.296 "name": "raid_bdev1", 00:12:53.296 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:12:53.296 "strip_size_kb": 0, 00:12:53.296 "state": "online", 00:12:53.296 "raid_level": "raid1", 00:12:53.296 "superblock": true, 00:12:53.296 "num_base_bdevs": 4, 00:12:53.296 "num_base_bdevs_discovered": 4, 00:12:53.296 "num_base_bdevs_operational": 4, 00:12:53.297 "base_bdevs_list": [ 00:12:53.297 { 00:12:53.297 "name": "BaseBdev1", 00:12:53.297 "uuid": "7bb16a54-e545-5cf5-a8ef-57f4a4fbb833", 00:12:53.297 "is_configured": true, 00:12:53.297 "data_offset": 2048, 00:12:53.297 "data_size": 63488 00:12:53.297 }, 00:12:53.297 { 00:12:53.297 "name": "BaseBdev2", 00:12:53.297 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:12:53.297 "is_configured": true, 00:12:53.297 "data_offset": 2048, 00:12:53.297 "data_size": 63488 00:12:53.297 }, 00:12:53.297 { 00:12:53.297 "name": "BaseBdev3", 00:12:53.297 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:12:53.297 "is_configured": true, 00:12:53.297 "data_offset": 2048, 00:12:53.297 "data_size": 63488 00:12:53.297 }, 00:12:53.297 { 00:12:53.297 "name": "BaseBdev4", 00:12:53.297 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:12:53.297 "is_configured": true, 00:12:53.297 "data_offset": 2048, 00:12:53.297 "data_size": 63488 00:12:53.297 } 00:12:53.297 ] 00:12:53.297 }' 00:12:53.297 11:01:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.297 11:01:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:53.866 [2024-10-29 11:01:59.161225] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:53.866 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:54.125 [2024-10-29 11:01:59.420554] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:12:54.125 /dev/nbd0 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:54.125 1+0 records in 00:12:54.125 1+0 records out 00:12:54.125 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000272367 s, 15.0 MB/s 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:54.125 11:01:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:59.401 63488+0 records in 00:12:59.401 63488+0 records out 00:12:59.401 32505856 bytes (33 MB, 31 MiB) copied, 4.74355 s, 6.9 MB/s 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:59.401 [2024-10-29 11:02:04.434751] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.401 [2024-10-29 11:02:04.442817] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.401 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.401 "name": "raid_bdev1", 00:12:59.401 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:12:59.401 "strip_size_kb": 0, 00:12:59.401 "state": "online", 00:12:59.401 "raid_level": "raid1", 00:12:59.401 "superblock": true, 00:12:59.401 "num_base_bdevs": 4, 00:12:59.401 "num_base_bdevs_discovered": 3, 00:12:59.401 "num_base_bdevs_operational": 3, 00:12:59.401 "base_bdevs_list": [ 00:12:59.401 { 00:12:59.401 "name": null, 00:12:59.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.402 "is_configured": false, 00:12:59.402 "data_offset": 0, 00:12:59.402 "data_size": 63488 00:12:59.402 }, 00:12:59.402 { 00:12:59.402 "name": "BaseBdev2", 00:12:59.402 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:12:59.402 "is_configured": true, 00:12:59.402 "data_offset": 2048, 00:12:59.402 "data_size": 63488 00:12:59.402 }, 00:12:59.402 { 00:12:59.402 "name": "BaseBdev3", 00:12:59.402 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:12:59.402 "is_configured": true, 00:12:59.402 "data_offset": 2048, 00:12:59.402 "data_size": 63488 00:12:59.402 }, 00:12:59.402 { 00:12:59.402 "name": "BaseBdev4", 00:12:59.402 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:12:59.402 "is_configured": true, 00:12:59.402 "data_offset": 2048, 00:12:59.402 "data_size": 63488 00:12:59.402 } 00:12:59.402 ] 00:12:59.402 }' 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:59.402 [2024-10-29 11:02:04.882138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:59.402 [2024-10-29 11:02:04.889590] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3360 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.402 11:02:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:59.402 [2024-10-29 11:02:04.891899] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:00.785 "name": "raid_bdev1", 00:13:00.785 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:00.785 "strip_size_kb": 0, 00:13:00.785 "state": "online", 00:13:00.785 "raid_level": "raid1", 00:13:00.785 "superblock": true, 00:13:00.785 "num_base_bdevs": 4, 00:13:00.785 "num_base_bdevs_discovered": 4, 00:13:00.785 "num_base_bdevs_operational": 4, 00:13:00.785 "process": { 00:13:00.785 "type": "rebuild", 00:13:00.785 "target": "spare", 00:13:00.785 "progress": { 00:13:00.785 "blocks": 20480, 00:13:00.785 "percent": 32 00:13:00.785 } 00:13:00.785 }, 00:13:00.785 "base_bdevs_list": [ 00:13:00.785 { 00:13:00.785 "name": "spare", 00:13:00.785 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev2", 00:13:00.785 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev3", 00:13:00.785 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev4", 00:13:00.785 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 } 00:13:00.785 ] 00:13:00.785 }' 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:00.785 11:02:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:00.785 [2024-10-29 11:02:06.052289] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:00.785 [2024-10-29 11:02:06.100188] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:00.785 [2024-10-29 11:02:06.100277] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.785 [2024-10-29 11:02:06.100305] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:00.785 [2024-10-29 11:02:06.100314] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.785 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:00.785 "name": "raid_bdev1", 00:13:00.785 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:00.785 "strip_size_kb": 0, 00:13:00.785 "state": "online", 00:13:00.785 "raid_level": "raid1", 00:13:00.785 "superblock": true, 00:13:00.785 "num_base_bdevs": 4, 00:13:00.785 "num_base_bdevs_discovered": 3, 00:13:00.785 "num_base_bdevs_operational": 3, 00:13:00.785 "base_bdevs_list": [ 00:13:00.785 { 00:13:00.785 "name": null, 00:13:00.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.785 "is_configured": false, 00:13:00.785 "data_offset": 0, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev2", 00:13:00.785 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev3", 00:13:00.785 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:00.785 "is_configured": true, 00:13:00.785 "data_offset": 2048, 00:13:00.785 "data_size": 63488 00:13:00.785 }, 00:13:00.785 { 00:13:00.785 "name": "BaseBdev4", 00:13:00.786 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:00.786 "is_configured": true, 00:13:00.786 "data_offset": 2048, 00:13:00.786 "data_size": 63488 00:13:00.786 } 00:13:00.786 ] 00:13:00.786 }' 00:13:00.786 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:00.786 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:01.355 "name": "raid_bdev1", 00:13:01.355 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:01.355 "strip_size_kb": 0, 00:13:01.355 "state": "online", 00:13:01.355 "raid_level": "raid1", 00:13:01.355 "superblock": true, 00:13:01.355 "num_base_bdevs": 4, 00:13:01.355 "num_base_bdevs_discovered": 3, 00:13:01.355 "num_base_bdevs_operational": 3, 00:13:01.355 "base_bdevs_list": [ 00:13:01.355 { 00:13:01.355 "name": null, 00:13:01.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.355 "is_configured": false, 00:13:01.355 "data_offset": 0, 00:13:01.355 "data_size": 63488 00:13:01.355 }, 00:13:01.355 { 00:13:01.355 "name": "BaseBdev2", 00:13:01.355 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:13:01.355 "is_configured": true, 00:13:01.355 "data_offset": 2048, 00:13:01.355 "data_size": 63488 00:13:01.355 }, 00:13:01.355 { 00:13:01.355 "name": "BaseBdev3", 00:13:01.355 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:01.355 "is_configured": true, 00:13:01.355 "data_offset": 2048, 00:13:01.355 "data_size": 63488 00:13:01.355 }, 00:13:01.355 { 00:13:01.355 "name": "BaseBdev4", 00:13:01.355 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:01.355 "is_configured": true, 00:13:01.355 "data_offset": 2048, 00:13:01.355 "data_size": 63488 00:13:01.355 } 00:13:01.355 ] 00:13:01.355 }' 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.355 [2024-10-29 11:02:06.720039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:01.355 [2024-10-29 11:02:06.724152] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.355 11:02:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:01.355 [2024-10-29 11:02:06.726161] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.294 "name": "raid_bdev1", 00:13:02.294 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:02.294 "strip_size_kb": 0, 00:13:02.294 "state": "online", 00:13:02.294 "raid_level": "raid1", 00:13:02.294 "superblock": true, 00:13:02.294 "num_base_bdevs": 4, 00:13:02.294 "num_base_bdevs_discovered": 4, 00:13:02.294 "num_base_bdevs_operational": 4, 00:13:02.294 "process": { 00:13:02.294 "type": "rebuild", 00:13:02.294 "target": "spare", 00:13:02.294 "progress": { 00:13:02.294 "blocks": 20480, 00:13:02.294 "percent": 32 00:13:02.294 } 00:13:02.294 }, 00:13:02.294 "base_bdevs_list": [ 00:13:02.294 { 00:13:02.294 "name": "spare", 00:13:02.294 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:02.294 "is_configured": true, 00:13:02.294 "data_offset": 2048, 00:13:02.294 "data_size": 63488 00:13:02.294 }, 00:13:02.294 { 00:13:02.294 "name": "BaseBdev2", 00:13:02.294 "uuid": "0af028ee-c616-5fe3-a547-ddcb4b8d431b", 00:13:02.294 "is_configured": true, 00:13:02.294 "data_offset": 2048, 00:13:02.294 "data_size": 63488 00:13:02.294 }, 00:13:02.294 { 00:13:02.294 "name": "BaseBdev3", 00:13:02.294 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:02.294 "is_configured": true, 00:13:02.294 "data_offset": 2048, 00:13:02.294 "data_size": 63488 00:13:02.294 }, 00:13:02.294 { 00:13:02.294 "name": "BaseBdev4", 00:13:02.294 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:02.294 "is_configured": true, 00:13:02.294 "data_offset": 2048, 00:13:02.294 "data_size": 63488 00:13:02.294 } 00:13:02.294 ] 00:13:02.294 }' 00:13:02.294 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:02.554 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.554 11:02:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.554 [2024-10-29 11:02:07.887241] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:02.554 [2024-10-29 11:02:08.030686] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca3430 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.554 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.816 "name": "raid_bdev1", 00:13:02.816 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:02.816 "strip_size_kb": 0, 00:13:02.816 "state": "online", 00:13:02.816 "raid_level": "raid1", 00:13:02.816 "superblock": true, 00:13:02.816 "num_base_bdevs": 4, 00:13:02.816 "num_base_bdevs_discovered": 3, 00:13:02.816 "num_base_bdevs_operational": 3, 00:13:02.816 "process": { 00:13:02.816 "type": "rebuild", 00:13:02.816 "target": "spare", 00:13:02.816 "progress": { 00:13:02.816 "blocks": 24576, 00:13:02.816 "percent": 38 00:13:02.816 } 00:13:02.816 }, 00:13:02.816 "base_bdevs_list": [ 00:13:02.816 { 00:13:02.816 "name": "spare", 00:13:02.816 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": null, 00:13:02.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.816 "is_configured": false, 00:13:02.816 "data_offset": 0, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": "BaseBdev3", 00:13:02.816 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": "BaseBdev4", 00:13:02.816 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 } 00:13:02.816 ] 00:13:02.816 }' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=373 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.816 "name": "raid_bdev1", 00:13:02.816 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:02.816 "strip_size_kb": 0, 00:13:02.816 "state": "online", 00:13:02.816 "raid_level": "raid1", 00:13:02.816 "superblock": true, 00:13:02.816 "num_base_bdevs": 4, 00:13:02.816 "num_base_bdevs_discovered": 3, 00:13:02.816 "num_base_bdevs_operational": 3, 00:13:02.816 "process": { 00:13:02.816 "type": "rebuild", 00:13:02.816 "target": "spare", 00:13:02.816 "progress": { 00:13:02.816 "blocks": 26624, 00:13:02.816 "percent": 41 00:13:02.816 } 00:13:02.816 }, 00:13:02.816 "base_bdevs_list": [ 00:13:02.816 { 00:13:02.816 "name": "spare", 00:13:02.816 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": null, 00:13:02.816 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.816 "is_configured": false, 00:13:02.816 "data_offset": 0, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": "BaseBdev3", 00:13:02.816 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 }, 00:13:02.816 { 00:13:02.816 "name": "BaseBdev4", 00:13:02.816 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:02.816 "is_configured": true, 00:13:02.816 "data_offset": 2048, 00:13:02.816 "data_size": 63488 00:13:02.816 } 00:13:02.816 ] 00:13:02.816 }' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.816 11:02:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.198 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.198 "name": "raid_bdev1", 00:13:04.198 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:04.198 "strip_size_kb": 0, 00:13:04.198 "state": "online", 00:13:04.198 "raid_level": "raid1", 00:13:04.198 "superblock": true, 00:13:04.198 "num_base_bdevs": 4, 00:13:04.198 "num_base_bdevs_discovered": 3, 00:13:04.198 "num_base_bdevs_operational": 3, 00:13:04.198 "process": { 00:13:04.198 "type": "rebuild", 00:13:04.198 "target": "spare", 00:13:04.198 "progress": { 00:13:04.198 "blocks": 49152, 00:13:04.198 "percent": 77 00:13:04.198 } 00:13:04.198 }, 00:13:04.198 "base_bdevs_list": [ 00:13:04.198 { 00:13:04.198 "name": "spare", 00:13:04.198 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:04.198 "is_configured": true, 00:13:04.198 "data_offset": 2048, 00:13:04.198 "data_size": 63488 00:13:04.198 }, 00:13:04.198 { 00:13:04.198 "name": null, 00:13:04.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.198 "is_configured": false, 00:13:04.198 "data_offset": 0, 00:13:04.199 "data_size": 63488 00:13:04.199 }, 00:13:04.199 { 00:13:04.199 "name": "BaseBdev3", 00:13:04.199 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:04.199 "is_configured": true, 00:13:04.199 "data_offset": 2048, 00:13:04.199 "data_size": 63488 00:13:04.199 }, 00:13:04.199 { 00:13:04.199 "name": "BaseBdev4", 00:13:04.199 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:04.199 "is_configured": true, 00:13:04.199 "data_offset": 2048, 00:13:04.199 "data_size": 63488 00:13:04.199 } 00:13:04.199 ] 00:13:04.199 }' 00:13:04.199 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.199 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.199 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.199 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:04.199 11:02:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:04.458 [2024-10-29 11:02:09.937496] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:04.458 [2024-10-29 11:02:09.937651] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:04.458 [2024-10-29 11:02:09.937822] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.027 "name": "raid_bdev1", 00:13:05.027 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:05.027 "strip_size_kb": 0, 00:13:05.027 "state": "online", 00:13:05.027 "raid_level": "raid1", 00:13:05.027 "superblock": true, 00:13:05.027 "num_base_bdevs": 4, 00:13:05.027 "num_base_bdevs_discovered": 3, 00:13:05.027 "num_base_bdevs_operational": 3, 00:13:05.027 "base_bdevs_list": [ 00:13:05.027 { 00:13:05.027 "name": "spare", 00:13:05.027 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:05.027 "is_configured": true, 00:13:05.027 "data_offset": 2048, 00:13:05.027 "data_size": 63488 00:13:05.027 }, 00:13:05.027 { 00:13:05.027 "name": null, 00:13:05.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.027 "is_configured": false, 00:13:05.027 "data_offset": 0, 00:13:05.027 "data_size": 63488 00:13:05.027 }, 00:13:05.027 { 00:13:05.027 "name": "BaseBdev3", 00:13:05.027 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:05.027 "is_configured": true, 00:13:05.027 "data_offset": 2048, 00:13:05.027 "data_size": 63488 00:13:05.027 }, 00:13:05.027 { 00:13:05.027 "name": "BaseBdev4", 00:13:05.027 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:05.027 "is_configured": true, 00:13:05.027 "data_offset": 2048, 00:13:05.027 "data_size": 63488 00:13:05.027 } 00:13:05.027 ] 00:13:05.027 }' 00:13:05.027 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.288 "name": "raid_bdev1", 00:13:05.288 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:05.288 "strip_size_kb": 0, 00:13:05.288 "state": "online", 00:13:05.288 "raid_level": "raid1", 00:13:05.288 "superblock": true, 00:13:05.288 "num_base_bdevs": 4, 00:13:05.288 "num_base_bdevs_discovered": 3, 00:13:05.288 "num_base_bdevs_operational": 3, 00:13:05.288 "base_bdevs_list": [ 00:13:05.288 { 00:13:05.288 "name": "spare", 00:13:05.288 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": null, 00:13:05.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.288 "is_configured": false, 00:13:05.288 "data_offset": 0, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": "BaseBdev3", 00:13:05.288 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": "BaseBdev4", 00:13:05.288 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 } 00:13:05.288 ] 00:13:05.288 }' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:05.288 "name": "raid_bdev1", 00:13:05.288 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:05.288 "strip_size_kb": 0, 00:13:05.288 "state": "online", 00:13:05.288 "raid_level": "raid1", 00:13:05.288 "superblock": true, 00:13:05.288 "num_base_bdevs": 4, 00:13:05.288 "num_base_bdevs_discovered": 3, 00:13:05.288 "num_base_bdevs_operational": 3, 00:13:05.288 "base_bdevs_list": [ 00:13:05.288 { 00:13:05.288 "name": "spare", 00:13:05.288 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": null, 00:13:05.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.288 "is_configured": false, 00:13:05.288 "data_offset": 0, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": "BaseBdev3", 00:13:05.288 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 }, 00:13:05.288 { 00:13:05.288 "name": "BaseBdev4", 00:13:05.288 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:05.288 "is_configured": true, 00:13:05.288 "data_offset": 2048, 00:13:05.288 "data_size": 63488 00:13:05.288 } 00:13:05.288 ] 00:13:05.288 }' 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:05.288 11:02:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.859 [2024-10-29 11:02:11.143817] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:05.859 [2024-10-29 11:02:11.143848] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:05.859 [2024-10-29 11:02:11.143952] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:05.859 [2024-10-29 11:02:11.144035] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:05.859 [2024-10-29 11:02:11.144058] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:05.859 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:06.119 /dev/nbd0 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:06.119 1+0 records in 00:13:06.119 1+0 records out 00:13:06.119 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000462262 s, 8.9 MB/s 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:06.119 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:06.379 /dev/nbd1 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:06.379 1+0 records in 00:13:06.379 1+0 records out 00:13:06.379 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000432773 s, 9.5 MB/s 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:06.379 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:06.638 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:06.638 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:06.638 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:06.639 11:02:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:06.899 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:06.899 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:06.899 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:06.900 [2024-10-29 11:02:12.207902] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:06.900 [2024-10-29 11:02:12.207984] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:06.900 [2024-10-29 11:02:12.208017] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:13:06.900 [2024-10-29 11:02:12.208036] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:06.900 [2024-10-29 11:02:12.210420] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:06.900 [2024-10-29 11:02:12.210513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:06.900 [2024-10-29 11:02:12.210645] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:06.900 [2024-10-29 11:02:12.210695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:06.900 [2024-10-29 11:02:12.210811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:06.900 [2024-10-29 11:02:12.210900] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:06.900 spare 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:06.900 [2024-10-29 11:02:12.310784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:13:06.900 [2024-10-29 11:02:12.310815] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:06.900 [2024-10-29 11:02:12.311098] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:13:06.900 [2024-10-29 11:02:12.311248] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:13:06.900 [2024-10-29 11:02:12.311258] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:13:06.900 [2024-10-29 11:02:12.311405] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.900 "name": "raid_bdev1", 00:13:06.900 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:06.900 "strip_size_kb": 0, 00:13:06.900 "state": "online", 00:13:06.900 "raid_level": "raid1", 00:13:06.900 "superblock": true, 00:13:06.900 "num_base_bdevs": 4, 00:13:06.900 "num_base_bdevs_discovered": 3, 00:13:06.900 "num_base_bdevs_operational": 3, 00:13:06.900 "base_bdevs_list": [ 00:13:06.900 { 00:13:06.900 "name": "spare", 00:13:06.900 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:06.900 "is_configured": true, 00:13:06.900 "data_offset": 2048, 00:13:06.900 "data_size": 63488 00:13:06.900 }, 00:13:06.900 { 00:13:06.900 "name": null, 00:13:06.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.900 "is_configured": false, 00:13:06.900 "data_offset": 2048, 00:13:06.900 "data_size": 63488 00:13:06.900 }, 00:13:06.900 { 00:13:06.900 "name": "BaseBdev3", 00:13:06.900 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:06.900 "is_configured": true, 00:13:06.900 "data_offset": 2048, 00:13:06.900 "data_size": 63488 00:13:06.900 }, 00:13:06.900 { 00:13:06.900 "name": "BaseBdev4", 00:13:06.900 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:06.900 "is_configured": true, 00:13:06.900 "data_offset": 2048, 00:13:06.900 "data_size": 63488 00:13:06.900 } 00:13:06.900 ] 00:13:06.900 }' 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.900 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.471 "name": "raid_bdev1", 00:13:07.471 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:07.471 "strip_size_kb": 0, 00:13:07.471 "state": "online", 00:13:07.471 "raid_level": "raid1", 00:13:07.471 "superblock": true, 00:13:07.471 "num_base_bdevs": 4, 00:13:07.471 "num_base_bdevs_discovered": 3, 00:13:07.471 "num_base_bdevs_operational": 3, 00:13:07.471 "base_bdevs_list": [ 00:13:07.471 { 00:13:07.471 "name": "spare", 00:13:07.471 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:07.471 "is_configured": true, 00:13:07.471 "data_offset": 2048, 00:13:07.471 "data_size": 63488 00:13:07.471 }, 00:13:07.471 { 00:13:07.471 "name": null, 00:13:07.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.471 "is_configured": false, 00:13:07.471 "data_offset": 2048, 00:13:07.471 "data_size": 63488 00:13:07.471 }, 00:13:07.471 { 00:13:07.471 "name": "BaseBdev3", 00:13:07.471 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:07.471 "is_configured": true, 00:13:07.471 "data_offset": 2048, 00:13:07.471 "data_size": 63488 00:13:07.471 }, 00:13:07.471 { 00:13:07.471 "name": "BaseBdev4", 00:13:07.471 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:07.471 "is_configured": true, 00:13:07.471 "data_offset": 2048, 00:13:07.471 "data_size": 63488 00:13:07.471 } 00:13:07.471 ] 00:13:07.471 }' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.471 [2024-10-29 11:02:12.926718] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.471 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.766 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.766 "name": "raid_bdev1", 00:13:07.766 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:07.766 "strip_size_kb": 0, 00:13:07.766 "state": "online", 00:13:07.766 "raid_level": "raid1", 00:13:07.766 "superblock": true, 00:13:07.766 "num_base_bdevs": 4, 00:13:07.766 "num_base_bdevs_discovered": 2, 00:13:07.766 "num_base_bdevs_operational": 2, 00:13:07.766 "base_bdevs_list": [ 00:13:07.766 { 00:13:07.766 "name": null, 00:13:07.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.766 "is_configured": false, 00:13:07.766 "data_offset": 0, 00:13:07.766 "data_size": 63488 00:13:07.766 }, 00:13:07.766 { 00:13:07.766 "name": null, 00:13:07.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.766 "is_configured": false, 00:13:07.766 "data_offset": 2048, 00:13:07.766 "data_size": 63488 00:13:07.766 }, 00:13:07.766 { 00:13:07.766 "name": "BaseBdev3", 00:13:07.766 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:07.766 "is_configured": true, 00:13:07.766 "data_offset": 2048, 00:13:07.766 "data_size": 63488 00:13:07.766 }, 00:13:07.766 { 00:13:07.766 "name": "BaseBdev4", 00:13:07.766 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:07.766 "is_configured": true, 00:13:07.766 "data_offset": 2048, 00:13:07.766 "data_size": 63488 00:13:07.766 } 00:13:07.766 ] 00:13:07.766 }' 00:13:07.766 11:02:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.766 11:02:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.026 11:02:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:08.026 11:02:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.026 11:02:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.026 [2024-10-29 11:02:13.385975] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:08.026 [2024-10-29 11:02:13.386228] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:08.026 [2024-10-29 11:02:13.386330] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:08.026 [2024-10-29 11:02:13.386747] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:08.026 [2024-10-29 11:02:13.391045] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:13:08.026 11:02:13 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.026 11:02:13 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:08.026 [2024-10-29 11:02:13.393136] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.966 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:08.966 "name": "raid_bdev1", 00:13:08.966 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:08.966 "strip_size_kb": 0, 00:13:08.966 "state": "online", 00:13:08.966 "raid_level": "raid1", 00:13:08.966 "superblock": true, 00:13:08.966 "num_base_bdevs": 4, 00:13:08.966 "num_base_bdevs_discovered": 3, 00:13:08.966 "num_base_bdevs_operational": 3, 00:13:08.966 "process": { 00:13:08.966 "type": "rebuild", 00:13:08.966 "target": "spare", 00:13:08.966 "progress": { 00:13:08.966 "blocks": 20480, 00:13:08.966 "percent": 32 00:13:08.966 } 00:13:08.966 }, 00:13:08.966 "base_bdevs_list": [ 00:13:08.966 { 00:13:08.966 "name": "spare", 00:13:08.966 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:08.966 "is_configured": true, 00:13:08.966 "data_offset": 2048, 00:13:08.966 "data_size": 63488 00:13:08.966 }, 00:13:08.966 { 00:13:08.966 "name": null, 00:13:08.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.966 "is_configured": false, 00:13:08.966 "data_offset": 2048, 00:13:08.966 "data_size": 63488 00:13:08.966 }, 00:13:08.966 { 00:13:08.966 "name": "BaseBdev3", 00:13:08.966 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:08.966 "is_configured": true, 00:13:08.966 "data_offset": 2048, 00:13:08.966 "data_size": 63488 00:13:08.966 }, 00:13:08.966 { 00:13:08.967 "name": "BaseBdev4", 00:13:08.967 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:08.967 "is_configured": true, 00:13:08.967 "data_offset": 2048, 00:13:08.967 "data_size": 63488 00:13:08.967 } 00:13:08.967 ] 00:13:08.967 }' 00:13:08.967 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.227 [2024-10-29 11:02:14.533087] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:09.227 [2024-10-29 11:02:14.597408] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:09.227 [2024-10-29 11:02:14.597880] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.227 [2024-10-29 11:02:14.597902] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:09.227 [2024-10-29 11:02:14.597915] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.227 "name": "raid_bdev1", 00:13:09.227 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:09.227 "strip_size_kb": 0, 00:13:09.227 "state": "online", 00:13:09.227 "raid_level": "raid1", 00:13:09.227 "superblock": true, 00:13:09.227 "num_base_bdevs": 4, 00:13:09.227 "num_base_bdevs_discovered": 2, 00:13:09.227 "num_base_bdevs_operational": 2, 00:13:09.227 "base_bdevs_list": [ 00:13:09.227 { 00:13:09.227 "name": null, 00:13:09.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.227 "is_configured": false, 00:13:09.227 "data_offset": 0, 00:13:09.227 "data_size": 63488 00:13:09.227 }, 00:13:09.227 { 00:13:09.227 "name": null, 00:13:09.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.227 "is_configured": false, 00:13:09.227 "data_offset": 2048, 00:13:09.227 "data_size": 63488 00:13:09.227 }, 00:13:09.227 { 00:13:09.227 "name": "BaseBdev3", 00:13:09.227 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:09.227 "is_configured": true, 00:13:09.227 "data_offset": 2048, 00:13:09.227 "data_size": 63488 00:13:09.227 }, 00:13:09.227 { 00:13:09.227 "name": "BaseBdev4", 00:13:09.227 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:09.227 "is_configured": true, 00:13:09.227 "data_offset": 2048, 00:13:09.227 "data_size": 63488 00:13:09.227 } 00:13:09.227 ] 00:13:09.227 }' 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.227 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.798 11:02:14 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:09.798 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.798 11:02:14 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.798 [2024-10-29 11:02:14.997477] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:09.798 [2024-10-29 11:02:14.997604] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.798 [2024-10-29 11:02:14.997667] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:13:09.798 [2024-10-29 11:02:14.997705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.798 [2024-10-29 11:02:14.998193] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.798 [2024-10-29 11:02:14.998275] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:09.798 [2024-10-29 11:02:14.998425] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:09.798 [2024-10-29 11:02:14.998485] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:09.798 [2024-10-29 11:02:14.998549] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:09.798 [2024-10-29 11:02:14.998616] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:09.798 [2024-10-29 11:02:15.002783] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:13:09.798 spare 00:13:09.798 11:02:15 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.798 11:02:15 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:09.798 [2024-10-29 11:02:15.004793] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:10.739 "name": "raid_bdev1", 00:13:10.739 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:10.739 "strip_size_kb": 0, 00:13:10.739 "state": "online", 00:13:10.739 "raid_level": "raid1", 00:13:10.739 "superblock": true, 00:13:10.739 "num_base_bdevs": 4, 00:13:10.739 "num_base_bdevs_discovered": 3, 00:13:10.739 "num_base_bdevs_operational": 3, 00:13:10.739 "process": { 00:13:10.739 "type": "rebuild", 00:13:10.739 "target": "spare", 00:13:10.739 "progress": { 00:13:10.739 "blocks": 20480, 00:13:10.739 "percent": 32 00:13:10.739 } 00:13:10.739 }, 00:13:10.739 "base_bdevs_list": [ 00:13:10.739 { 00:13:10.739 "name": "spare", 00:13:10.739 "uuid": "9eaf9981-b61a-55ad-8674-d1fa461015f4", 00:13:10.739 "is_configured": true, 00:13:10.739 "data_offset": 2048, 00:13:10.739 "data_size": 63488 00:13:10.739 }, 00:13:10.739 { 00:13:10.739 "name": null, 00:13:10.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.739 "is_configured": false, 00:13:10.739 "data_offset": 2048, 00:13:10.739 "data_size": 63488 00:13:10.739 }, 00:13:10.739 { 00:13:10.739 "name": "BaseBdev3", 00:13:10.739 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:10.739 "is_configured": true, 00:13:10.739 "data_offset": 2048, 00:13:10.739 "data_size": 63488 00:13:10.739 }, 00:13:10.739 { 00:13:10.739 "name": "BaseBdev4", 00:13:10.739 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:10.739 "is_configured": true, 00:13:10.739 "data_offset": 2048, 00:13:10.739 "data_size": 63488 00:13:10.739 } 00:13:10.739 ] 00:13:10.739 }' 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:10.739 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.740 [2024-10-29 11:02:16.169078] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:10.740 [2024-10-29 11:02:16.208974] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:10.740 [2024-10-29 11:02:16.209031] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:10.740 [2024-10-29 11:02:16.209051] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:10.740 [2024-10-29 11:02:16.209060] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.740 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.000 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.000 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.000 "name": "raid_bdev1", 00:13:11.000 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:11.000 "strip_size_kb": 0, 00:13:11.000 "state": "online", 00:13:11.000 "raid_level": "raid1", 00:13:11.000 "superblock": true, 00:13:11.000 "num_base_bdevs": 4, 00:13:11.000 "num_base_bdevs_discovered": 2, 00:13:11.000 "num_base_bdevs_operational": 2, 00:13:11.000 "base_bdevs_list": [ 00:13:11.000 { 00:13:11.000 "name": null, 00:13:11.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.000 "is_configured": false, 00:13:11.000 "data_offset": 0, 00:13:11.000 "data_size": 63488 00:13:11.000 }, 00:13:11.000 { 00:13:11.000 "name": null, 00:13:11.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.000 "is_configured": false, 00:13:11.000 "data_offset": 2048, 00:13:11.000 "data_size": 63488 00:13:11.000 }, 00:13:11.000 { 00:13:11.000 "name": "BaseBdev3", 00:13:11.000 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:11.000 "is_configured": true, 00:13:11.000 "data_offset": 2048, 00:13:11.000 "data_size": 63488 00:13:11.000 }, 00:13:11.000 { 00:13:11.000 "name": "BaseBdev4", 00:13:11.000 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:11.000 "is_configured": true, 00:13:11.000 "data_offset": 2048, 00:13:11.000 "data_size": 63488 00:13:11.000 } 00:13:11.000 ] 00:13:11.000 }' 00:13:11.000 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.000 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.260 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.261 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.261 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:11.261 "name": "raid_bdev1", 00:13:11.261 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:11.261 "strip_size_kb": 0, 00:13:11.261 "state": "online", 00:13:11.261 "raid_level": "raid1", 00:13:11.261 "superblock": true, 00:13:11.261 "num_base_bdevs": 4, 00:13:11.261 "num_base_bdevs_discovered": 2, 00:13:11.261 "num_base_bdevs_operational": 2, 00:13:11.261 "base_bdevs_list": [ 00:13:11.261 { 00:13:11.261 "name": null, 00:13:11.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.261 "is_configured": false, 00:13:11.261 "data_offset": 0, 00:13:11.261 "data_size": 63488 00:13:11.261 }, 00:13:11.261 { 00:13:11.261 "name": null, 00:13:11.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.261 "is_configured": false, 00:13:11.261 "data_offset": 2048, 00:13:11.261 "data_size": 63488 00:13:11.261 }, 00:13:11.261 { 00:13:11.261 "name": "BaseBdev3", 00:13:11.261 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:11.261 "is_configured": true, 00:13:11.261 "data_offset": 2048, 00:13:11.261 "data_size": 63488 00:13:11.261 }, 00:13:11.261 { 00:13:11.261 "name": "BaseBdev4", 00:13:11.261 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:11.261 "is_configured": true, 00:13:11.261 "data_offset": 2048, 00:13:11.261 "data_size": 63488 00:13:11.261 } 00:13:11.261 ] 00:13:11.261 }' 00:13:11.261 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:11.261 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:11.261 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.521 [2024-10-29 11:02:16.796246] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:11.521 [2024-10-29 11:02:16.796307] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:11.521 [2024-10-29 11:02:16.796331] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:13:11.521 [2024-10-29 11:02:16.796342] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:11.521 [2024-10-29 11:02:16.796851] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:11.521 [2024-10-29 11:02:16.796892] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:11.521 [2024-10-29 11:02:16.796970] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:11.521 [2024-10-29 11:02:16.796983] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:11.521 [2024-10-29 11:02:16.797005] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:11.521 [2024-10-29 11:02:16.797016] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:11.521 BaseBdev1 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.521 11:02:16 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.462 "name": "raid_bdev1", 00:13:12.462 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:12.462 "strip_size_kb": 0, 00:13:12.462 "state": "online", 00:13:12.462 "raid_level": "raid1", 00:13:12.462 "superblock": true, 00:13:12.462 "num_base_bdevs": 4, 00:13:12.462 "num_base_bdevs_discovered": 2, 00:13:12.462 "num_base_bdevs_operational": 2, 00:13:12.462 "base_bdevs_list": [ 00:13:12.462 { 00:13:12.462 "name": null, 00:13:12.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.462 "is_configured": false, 00:13:12.462 "data_offset": 0, 00:13:12.462 "data_size": 63488 00:13:12.462 }, 00:13:12.462 { 00:13:12.462 "name": null, 00:13:12.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.462 "is_configured": false, 00:13:12.462 "data_offset": 2048, 00:13:12.462 "data_size": 63488 00:13:12.462 }, 00:13:12.462 { 00:13:12.462 "name": "BaseBdev3", 00:13:12.462 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:12.462 "is_configured": true, 00:13:12.462 "data_offset": 2048, 00:13:12.462 "data_size": 63488 00:13:12.462 }, 00:13:12.462 { 00:13:12.462 "name": "BaseBdev4", 00:13:12.462 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:12.462 "is_configured": true, 00:13:12.462 "data_offset": 2048, 00:13:12.462 "data_size": 63488 00:13:12.462 } 00:13:12.462 ] 00:13:12.462 }' 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.462 11:02:17 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:13.034 "name": "raid_bdev1", 00:13:13.034 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:13.034 "strip_size_kb": 0, 00:13:13.034 "state": "online", 00:13:13.034 "raid_level": "raid1", 00:13:13.034 "superblock": true, 00:13:13.034 "num_base_bdevs": 4, 00:13:13.034 "num_base_bdevs_discovered": 2, 00:13:13.034 "num_base_bdevs_operational": 2, 00:13:13.034 "base_bdevs_list": [ 00:13:13.034 { 00:13:13.034 "name": null, 00:13:13.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.034 "is_configured": false, 00:13:13.034 "data_offset": 0, 00:13:13.034 "data_size": 63488 00:13:13.034 }, 00:13:13.034 { 00:13:13.034 "name": null, 00:13:13.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.034 "is_configured": false, 00:13:13.034 "data_offset": 2048, 00:13:13.034 "data_size": 63488 00:13:13.034 }, 00:13:13.034 { 00:13:13.034 "name": "BaseBdev3", 00:13:13.034 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:13.034 "is_configured": true, 00:13:13.034 "data_offset": 2048, 00:13:13.034 "data_size": 63488 00:13:13.034 }, 00:13:13.034 { 00:13:13.034 "name": "BaseBdev4", 00:13:13.034 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:13.034 "is_configured": true, 00:13:13.034 "data_offset": 2048, 00:13:13.034 "data_size": 63488 00:13:13.034 } 00:13:13.034 ] 00:13:13.034 }' 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.034 [2024-10-29 11:02:18.365807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:13.034 [2024-10-29 11:02:18.365969] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:13.034 [2024-10-29 11:02:18.365989] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:13.034 request: 00:13:13.034 { 00:13:13.034 "base_bdev": "BaseBdev1", 00:13:13.034 "raid_bdev": "raid_bdev1", 00:13:13.034 "method": "bdev_raid_add_base_bdev", 00:13:13.034 "req_id": 1 00:13:13.034 } 00:13:13.034 Got JSON-RPC error response 00:13:13.034 response: 00:13:13.034 { 00:13:13.034 "code": -22, 00:13:13.034 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:13.034 } 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:13.034 11:02:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.974 "name": "raid_bdev1", 00:13:13.974 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:13.974 "strip_size_kb": 0, 00:13:13.974 "state": "online", 00:13:13.974 "raid_level": "raid1", 00:13:13.974 "superblock": true, 00:13:13.974 "num_base_bdevs": 4, 00:13:13.974 "num_base_bdevs_discovered": 2, 00:13:13.974 "num_base_bdevs_operational": 2, 00:13:13.974 "base_bdevs_list": [ 00:13:13.974 { 00:13:13.974 "name": null, 00:13:13.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.974 "is_configured": false, 00:13:13.974 "data_offset": 0, 00:13:13.974 "data_size": 63488 00:13:13.974 }, 00:13:13.974 { 00:13:13.974 "name": null, 00:13:13.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.974 "is_configured": false, 00:13:13.974 "data_offset": 2048, 00:13:13.974 "data_size": 63488 00:13:13.974 }, 00:13:13.974 { 00:13:13.974 "name": "BaseBdev3", 00:13:13.974 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:13.974 "is_configured": true, 00:13:13.974 "data_offset": 2048, 00:13:13.974 "data_size": 63488 00:13:13.974 }, 00:13:13.974 { 00:13:13.974 "name": "BaseBdev4", 00:13:13.974 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:13.974 "is_configured": true, 00:13:13.974 "data_offset": 2048, 00:13:13.974 "data_size": 63488 00:13:13.974 } 00:13:13.974 ] 00:13:13.974 }' 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.974 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.548 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:14.548 "name": "raid_bdev1", 00:13:14.548 "uuid": "eb3c6629-5b93-47c1-8354-acd60cd45cee", 00:13:14.548 "strip_size_kb": 0, 00:13:14.548 "state": "online", 00:13:14.548 "raid_level": "raid1", 00:13:14.548 "superblock": true, 00:13:14.548 "num_base_bdevs": 4, 00:13:14.548 "num_base_bdevs_discovered": 2, 00:13:14.548 "num_base_bdevs_operational": 2, 00:13:14.548 "base_bdevs_list": [ 00:13:14.548 { 00:13:14.548 "name": null, 00:13:14.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.548 "is_configured": false, 00:13:14.548 "data_offset": 0, 00:13:14.548 "data_size": 63488 00:13:14.548 }, 00:13:14.548 { 00:13:14.548 "name": null, 00:13:14.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.548 "is_configured": false, 00:13:14.548 "data_offset": 2048, 00:13:14.548 "data_size": 63488 00:13:14.548 }, 00:13:14.548 { 00:13:14.548 "name": "BaseBdev3", 00:13:14.548 "uuid": "9b0f73b6-f017-50b5-9049-281ef28c2c33", 00:13:14.548 "is_configured": true, 00:13:14.548 "data_offset": 2048, 00:13:14.548 "data_size": 63488 00:13:14.548 }, 00:13:14.548 { 00:13:14.548 "name": "BaseBdev4", 00:13:14.549 "uuid": "ab9dfe0a-b58f-539a-b0e5-dfdcdb49c29d", 00:13:14.549 "is_configured": true, 00:13:14.549 "data_offset": 2048, 00:13:14.549 "data_size": 63488 00:13:14.549 } 00:13:14.549 ] 00:13:14.549 }' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88737 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@952 -- # '[' -z 88737 ']' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # kill -0 88737 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@957 -- # uname 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 88737 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:13:14.549 killing process with pid 88737 00:13:14.549 Received shutdown signal, test time was about 60.000000 seconds 00:13:14.549 00:13:14.549 Latency(us) 00:13:14.549 [2024-10-29T11:02:20.047Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:14.549 [2024-10-29T11:02:20.047Z] =================================================================================================================== 00:13:14.549 [2024-10-29T11:02:20.047Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 88737' 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@971 -- # kill 88737 00:13:14.549 [2024-10-29 11:02:19.987757] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:14.549 [2024-10-29 11:02:19.987890] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:14.549 11:02:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@976 -- # wait 88737 00:13:14.549 [2024-10-29 11:02:19.987960] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:14.549 [2024-10-29 11:02:19.987973] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:13:14.549 [2024-10-29 11:02:20.038666] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:14.807 11:02:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:13:14.807 00:13:14.807 real 0m22.602s 00:13:14.807 user 0m27.733s 00:13:14.807 sys 0m3.559s 00:13:14.807 11:02:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:13:14.807 ************************************ 00:13:14.807 END TEST raid_rebuild_test_sb 00:13:14.807 ************************************ 00:13:14.807 11:02:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.807 11:02:20 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:13:14.807 11:02:20 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:13:14.807 11:02:20 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:13:14.807 11:02:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:15.083 ************************************ 00:13:15.083 START TEST raid_rebuild_test_io 00:13:15.083 ************************************ 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 4 false true true 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:15.083 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89468 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89468 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@833 -- # '[' -z 89468 ']' 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:15.084 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # local max_retries=100 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # xtrace_disable 00:13:15.084 11:02:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.084 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:15.084 Zero copy mechanism will not be used. 00:13:15.084 [2024-10-29 11:02:20.411302] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:13:15.084 [2024-10-29 11:02:20.411446] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89468 ] 00:13:15.365 [2024-10-29 11:02:20.581700] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:15.365 [2024-10-29 11:02:20.607445] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:15.365 [2024-10-29 11:02:20.650578] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:15.365 [2024-10-29 11:02:20.650620] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@866 -- # return 0 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.952 BaseBdev1_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.952 [2024-10-29 11:02:21.253935] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:15.952 [2024-10-29 11:02:21.254018] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.952 [2024-10-29 11:02:21.254048] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:15.952 [2024-10-29 11:02:21.254065] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.952 [2024-10-29 11:02:21.256237] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.952 [2024-10-29 11:02:21.256279] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:15.952 BaseBdev1 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.952 BaseBdev2_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.952 [2024-10-29 11:02:21.282628] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:15.952 [2024-10-29 11:02:21.282747] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.952 [2024-10-29 11:02:21.282774] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:15.952 [2024-10-29 11:02:21.282785] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.952 [2024-10-29 11:02:21.284919] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.952 [2024-10-29 11:02:21.284959] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:15.952 BaseBdev2 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.952 BaseBdev3_malloc 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:15.952 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 [2024-10-29 11:02:21.311717] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:15.953 [2024-10-29 11:02:21.311778] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.953 [2024-10-29 11:02:21.311805] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:15.953 [2024-10-29 11:02:21.311816] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.953 [2024-10-29 11:02:21.314085] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.953 [2024-10-29 11:02:21.314189] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:15.953 BaseBdev3 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 BaseBdev4_malloc 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 [2024-10-29 11:02:21.356609] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:15.953 [2024-10-29 11:02:21.356715] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.953 [2024-10-29 11:02:21.356768] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:15.953 [2024-10-29 11:02:21.356793] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.953 [2024-10-29 11:02:21.360350] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.953 [2024-10-29 11:02:21.360430] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:15.953 BaseBdev4 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 spare_malloc 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 spare_delay 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 [2024-10-29 11:02:21.397857] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:15.953 [2024-10-29 11:02:21.397916] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:15.953 [2024-10-29 11:02:21.397939] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:15.953 [2024-10-29 11:02:21.397949] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:15.953 [2024-10-29 11:02:21.400109] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:15.953 [2024-10-29 11:02:21.400207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:15.953 spare 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 [2024-10-29 11:02:21.405917] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:15.953 [2024-10-29 11:02:21.407862] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:15.953 [2024-10-29 11:02:21.407940] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:15.953 [2024-10-29 11:02:21.407989] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:15.953 [2024-10-29 11:02:21.408078] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:15.953 [2024-10-29 11:02:21.408088] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:15.953 [2024-10-29 11:02:21.408349] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:15.953 [2024-10-29 11:02:21.408531] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:15.953 [2024-10-29 11:02:21.408553] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:15.953 [2024-10-29 11:02:21.408666] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.953 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.232 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.232 "name": "raid_bdev1", 00:13:16.232 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:16.232 "strip_size_kb": 0, 00:13:16.232 "state": "online", 00:13:16.232 "raid_level": "raid1", 00:13:16.232 "superblock": false, 00:13:16.232 "num_base_bdevs": 4, 00:13:16.232 "num_base_bdevs_discovered": 4, 00:13:16.232 "num_base_bdevs_operational": 4, 00:13:16.232 "base_bdevs_list": [ 00:13:16.232 { 00:13:16.232 "name": "BaseBdev1", 00:13:16.232 "uuid": "ec3c4adf-9580-5010-b9d2-b686c3b85517", 00:13:16.232 "is_configured": true, 00:13:16.232 "data_offset": 0, 00:13:16.232 "data_size": 65536 00:13:16.232 }, 00:13:16.232 { 00:13:16.232 "name": "BaseBdev2", 00:13:16.232 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:16.232 "is_configured": true, 00:13:16.232 "data_offset": 0, 00:13:16.232 "data_size": 65536 00:13:16.232 }, 00:13:16.232 { 00:13:16.232 "name": "BaseBdev3", 00:13:16.232 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:16.232 "is_configured": true, 00:13:16.232 "data_offset": 0, 00:13:16.232 "data_size": 65536 00:13:16.232 }, 00:13:16.232 { 00:13:16.232 "name": "BaseBdev4", 00:13:16.232 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:16.232 "is_configured": true, 00:13:16.232 "data_offset": 0, 00:13:16.232 "data_size": 65536 00:13:16.232 } 00:13:16.232 ] 00:13:16.232 }' 00:13:16.232 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.232 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.505 [2024-10-29 11:02:21.861461] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.505 [2024-10-29 11:02:21.944968] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.505 "name": "raid_bdev1", 00:13:16.505 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:16.505 "strip_size_kb": 0, 00:13:16.505 "state": "online", 00:13:16.505 "raid_level": "raid1", 00:13:16.505 "superblock": false, 00:13:16.505 "num_base_bdevs": 4, 00:13:16.505 "num_base_bdevs_discovered": 3, 00:13:16.505 "num_base_bdevs_operational": 3, 00:13:16.505 "base_bdevs_list": [ 00:13:16.505 { 00:13:16.505 "name": null, 00:13:16.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.505 "is_configured": false, 00:13:16.505 "data_offset": 0, 00:13:16.505 "data_size": 65536 00:13:16.505 }, 00:13:16.505 { 00:13:16.505 "name": "BaseBdev2", 00:13:16.505 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:16.505 "is_configured": true, 00:13:16.505 "data_offset": 0, 00:13:16.505 "data_size": 65536 00:13:16.505 }, 00:13:16.505 { 00:13:16.505 "name": "BaseBdev3", 00:13:16.505 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:16.505 "is_configured": true, 00:13:16.505 "data_offset": 0, 00:13:16.505 "data_size": 65536 00:13:16.505 }, 00:13:16.505 { 00:13:16.505 "name": "BaseBdev4", 00:13:16.505 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:16.505 "is_configured": true, 00:13:16.505 "data_offset": 0, 00:13:16.505 "data_size": 65536 00:13:16.505 } 00:13:16.505 ] 00:13:16.505 }' 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.505 11:02:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.765 [2024-10-29 11:02:22.034904] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:16.765 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:16.765 Zero copy mechanism will not be used. 00:13:16.765 Running I/O for 60 seconds... 00:13:17.024 11:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:17.024 11:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.024 11:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.024 [2024-10-29 11:02:22.394589] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:17.024 11:02:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.024 11:02:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:17.024 [2024-10-29 11:02:22.449566] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:17.024 [2024-10-29 11:02:22.451847] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:17.283 [2024-10-29 11:02:22.575302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:17.283 [2024-10-29 11:02:22.575994] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:17.283 [2024-10-29 11:02:22.704130] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:17.283 [2024-10-29 11:02:22.704562] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:17.852 172.00 IOPS, 516.00 MiB/s [2024-10-29T11:02:23.350Z] [2024-10-29 11:02:23.044567] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:17.852 [2024-10-29 11:02:23.166452] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:17.852 [2024-10-29 11:02:23.166808] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.111 "name": "raid_bdev1", 00:13:18.111 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:18.111 "strip_size_kb": 0, 00:13:18.111 "state": "online", 00:13:18.111 "raid_level": "raid1", 00:13:18.111 "superblock": false, 00:13:18.111 "num_base_bdevs": 4, 00:13:18.111 "num_base_bdevs_discovered": 4, 00:13:18.111 "num_base_bdevs_operational": 4, 00:13:18.111 "process": { 00:13:18.111 "type": "rebuild", 00:13:18.111 "target": "spare", 00:13:18.111 "progress": { 00:13:18.111 "blocks": 12288, 00:13:18.111 "percent": 18 00:13:18.111 } 00:13:18.111 }, 00:13:18.111 "base_bdevs_list": [ 00:13:18.111 { 00:13:18.111 "name": "spare", 00:13:18.111 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:18.111 "is_configured": true, 00:13:18.111 "data_offset": 0, 00:13:18.111 "data_size": 65536 00:13:18.111 }, 00:13:18.111 { 00:13:18.111 "name": "BaseBdev2", 00:13:18.111 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:18.111 "is_configured": true, 00:13:18.111 "data_offset": 0, 00:13:18.111 "data_size": 65536 00:13:18.111 }, 00:13:18.111 { 00:13:18.111 "name": "BaseBdev3", 00:13:18.111 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:18.111 "is_configured": true, 00:13:18.111 "data_offset": 0, 00:13:18.111 "data_size": 65536 00:13:18.111 }, 00:13:18.111 { 00:13:18.111 "name": "BaseBdev4", 00:13:18.111 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:18.111 "is_configured": true, 00:13:18.111 "data_offset": 0, 00:13:18.111 "data_size": 65536 00:13:18.111 } 00:13:18.111 ] 00:13:18.111 }' 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.111 [2024-10-29 11:02:23.492527] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.111 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.111 [2024-10-29 11:02:23.600910] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:18.111 [2024-10-29 11:02:23.608682] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:18.371 [2024-10-29 11:02:23.608942] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:18.371 [2024-10-29 11:02:23.716527] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:18.371 [2024-10-29 11:02:23.727258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:18.371 [2024-10-29 11:02:23.727412] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:18.371 [2024-10-29 11:02:23.727447] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:18.371 [2024-10-29 11:02:23.740125] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:18.371 "name": "raid_bdev1", 00:13:18.371 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:18.371 "strip_size_kb": 0, 00:13:18.371 "state": "online", 00:13:18.371 "raid_level": "raid1", 00:13:18.371 "superblock": false, 00:13:18.371 "num_base_bdevs": 4, 00:13:18.371 "num_base_bdevs_discovered": 3, 00:13:18.371 "num_base_bdevs_operational": 3, 00:13:18.371 "base_bdevs_list": [ 00:13:18.371 { 00:13:18.371 "name": null, 00:13:18.371 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.371 "is_configured": false, 00:13:18.371 "data_offset": 0, 00:13:18.371 "data_size": 65536 00:13:18.371 }, 00:13:18.371 { 00:13:18.371 "name": "BaseBdev2", 00:13:18.371 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:18.371 "is_configured": true, 00:13:18.371 "data_offset": 0, 00:13:18.371 "data_size": 65536 00:13:18.371 }, 00:13:18.371 { 00:13:18.371 "name": "BaseBdev3", 00:13:18.371 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:18.371 "is_configured": true, 00:13:18.371 "data_offset": 0, 00:13:18.371 "data_size": 65536 00:13:18.371 }, 00:13:18.371 { 00:13:18.371 "name": "BaseBdev4", 00:13:18.371 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:18.371 "is_configured": true, 00:13:18.371 "data_offset": 0, 00:13:18.371 "data_size": 65536 00:13:18.371 } 00:13:18.371 ] 00:13:18.371 }' 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:18.371 11:02:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.890 146.50 IOPS, 439.50 MiB/s [2024-10-29T11:02:24.388Z] 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.890 "name": "raid_bdev1", 00:13:18.890 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:18.890 "strip_size_kb": 0, 00:13:18.890 "state": "online", 00:13:18.890 "raid_level": "raid1", 00:13:18.890 "superblock": false, 00:13:18.890 "num_base_bdevs": 4, 00:13:18.890 "num_base_bdevs_discovered": 3, 00:13:18.890 "num_base_bdevs_operational": 3, 00:13:18.890 "base_bdevs_list": [ 00:13:18.890 { 00:13:18.890 "name": null, 00:13:18.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.890 "is_configured": false, 00:13:18.890 "data_offset": 0, 00:13:18.890 "data_size": 65536 00:13:18.890 }, 00:13:18.890 { 00:13:18.890 "name": "BaseBdev2", 00:13:18.890 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:18.890 "is_configured": true, 00:13:18.890 "data_offset": 0, 00:13:18.890 "data_size": 65536 00:13:18.890 }, 00:13:18.890 { 00:13:18.890 "name": "BaseBdev3", 00:13:18.890 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:18.890 "is_configured": true, 00:13:18.890 "data_offset": 0, 00:13:18.890 "data_size": 65536 00:13:18.890 }, 00:13:18.890 { 00:13:18.890 "name": "BaseBdev4", 00:13:18.890 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:18.890 "is_configured": true, 00:13:18.890 "data_offset": 0, 00:13:18.890 "data_size": 65536 00:13:18.890 } 00:13:18.890 ] 00:13:18.890 }' 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.890 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.890 [2024-10-29 11:02:24.365137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:19.150 11:02:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.150 11:02:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:19.150 [2024-10-29 11:02:24.425320] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:19.150 [2024-10-29 11:02:24.427421] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:19.150 [2024-10-29 11:02:24.542354] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:19.150 [2024-10-29 11:02:24.543598] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:19.410 [2024-10-29 11:02:24.768026] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:19.410 [2024-10-29 11:02:24.768568] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:19.670 [2024-10-29 11:02:25.020466] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:19.670 148.00 IOPS, 444.00 MiB/s [2024-10-29T11:02:25.168Z] [2024-10-29 11:02:25.142973] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.931 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.191 "name": "raid_bdev1", 00:13:20.191 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:20.191 "strip_size_kb": 0, 00:13:20.191 "state": "online", 00:13:20.191 "raid_level": "raid1", 00:13:20.191 "superblock": false, 00:13:20.191 "num_base_bdevs": 4, 00:13:20.191 "num_base_bdevs_discovered": 4, 00:13:20.191 "num_base_bdevs_operational": 4, 00:13:20.191 "process": { 00:13:20.191 "type": "rebuild", 00:13:20.191 "target": "spare", 00:13:20.191 "progress": { 00:13:20.191 "blocks": 12288, 00:13:20.191 "percent": 18 00:13:20.191 } 00:13:20.191 }, 00:13:20.191 "base_bdevs_list": [ 00:13:20.191 { 00:13:20.191 "name": "spare", 00:13:20.191 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:20.191 "is_configured": true, 00:13:20.191 "data_offset": 0, 00:13:20.191 "data_size": 65536 00:13:20.191 }, 00:13:20.191 { 00:13:20.191 "name": "BaseBdev2", 00:13:20.191 "uuid": "960e8c4d-166c-5fc1-9c7c-583818ef2d61", 00:13:20.191 "is_configured": true, 00:13:20.191 "data_offset": 0, 00:13:20.191 "data_size": 65536 00:13:20.191 }, 00:13:20.191 { 00:13:20.191 "name": "BaseBdev3", 00:13:20.191 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:20.191 "is_configured": true, 00:13:20.191 "data_offset": 0, 00:13:20.191 "data_size": 65536 00:13:20.191 }, 00:13:20.191 { 00:13:20.191 "name": "BaseBdev4", 00:13:20.191 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:20.191 "is_configured": true, 00:13:20.191 "data_offset": 0, 00:13:20.191 "data_size": 65536 00:13:20.191 } 00:13:20.191 ] 00:13:20.191 }' 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.191 [2024-10-29 11:02:25.490317] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.191 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.191 [2024-10-29 11:02:25.526488] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:20.451 [2024-10-29 11:02:25.720637] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:20.451 [2024-10-29 11:02:25.726395] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:20.451 [2024-10-29 11:02:25.834287] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006080 00:13:20.451 [2024-10-29 11:02:25.834411] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:13:20.451 [2024-10-29 11:02:25.835271] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:20.451 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.451 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.452 "name": "raid_bdev1", 00:13:20.452 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:20.452 "strip_size_kb": 0, 00:13:20.452 "state": "online", 00:13:20.452 "raid_level": "raid1", 00:13:20.452 "superblock": false, 00:13:20.452 "num_base_bdevs": 4, 00:13:20.452 "num_base_bdevs_discovered": 3, 00:13:20.452 "num_base_bdevs_operational": 3, 00:13:20.452 "process": { 00:13:20.452 "type": "rebuild", 00:13:20.452 "target": "spare", 00:13:20.452 "progress": { 00:13:20.452 "blocks": 16384, 00:13:20.452 "percent": 25 00:13:20.452 } 00:13:20.452 }, 00:13:20.452 "base_bdevs_list": [ 00:13:20.452 { 00:13:20.452 "name": "spare", 00:13:20.452 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:20.452 "is_configured": true, 00:13:20.452 "data_offset": 0, 00:13:20.452 "data_size": 65536 00:13:20.452 }, 00:13:20.452 { 00:13:20.452 "name": null, 00:13:20.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.452 "is_configured": false, 00:13:20.452 "data_offset": 0, 00:13:20.452 "data_size": 65536 00:13:20.452 }, 00:13:20.452 { 00:13:20.452 "name": "BaseBdev3", 00:13:20.452 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:20.452 "is_configured": true, 00:13:20.452 "data_offset": 0, 00:13:20.452 "data_size": 65536 00:13:20.452 }, 00:13:20.452 { 00:13:20.452 "name": "BaseBdev4", 00:13:20.452 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:20.452 "is_configured": true, 00:13:20.452 "data_offset": 0, 00:13:20.452 "data_size": 65536 00:13:20.452 } 00:13:20.452 ] 00:13:20.452 }' 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:20.452 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=390 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.712 11:02:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.712 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.712 "name": "raid_bdev1", 00:13:20.712 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:20.712 "strip_size_kb": 0, 00:13:20.712 "state": "online", 00:13:20.712 "raid_level": "raid1", 00:13:20.712 "superblock": false, 00:13:20.712 "num_base_bdevs": 4, 00:13:20.712 "num_base_bdevs_discovered": 3, 00:13:20.712 "num_base_bdevs_operational": 3, 00:13:20.712 "process": { 00:13:20.712 "type": "rebuild", 00:13:20.712 "target": "spare", 00:13:20.712 "progress": { 00:13:20.712 "blocks": 16384, 00:13:20.712 "percent": 25 00:13:20.712 } 00:13:20.712 }, 00:13:20.712 "base_bdevs_list": [ 00:13:20.712 { 00:13:20.712 "name": "spare", 00:13:20.712 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:20.712 "is_configured": true, 00:13:20.712 "data_offset": 0, 00:13:20.712 "data_size": 65536 00:13:20.712 }, 00:13:20.712 { 00:13:20.712 "name": null, 00:13:20.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.712 "is_configured": false, 00:13:20.712 "data_offset": 0, 00:13:20.712 "data_size": 65536 00:13:20.712 }, 00:13:20.712 { 00:13:20.712 "name": "BaseBdev3", 00:13:20.712 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:20.712 "is_configured": true, 00:13:20.712 "data_offset": 0, 00:13:20.712 "data_size": 65536 00:13:20.713 }, 00:13:20.713 { 00:13:20.713 "name": "BaseBdev4", 00:13:20.713 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:20.713 "is_configured": true, 00:13:20.713 "data_offset": 0, 00:13:20.713 "data_size": 65536 00:13:20.713 } 00:13:20.713 ] 00:13:20.713 }' 00:13:20.713 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:20.713 127.00 IOPS, 381.00 MiB/s [2024-10-29T11:02:26.211Z] 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:20.713 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:20.713 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:20.713 11:02:26 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:20.973 [2024-10-29 11:02:26.281051] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:21.233 [2024-10-29 11:02:26.613334] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:21.493 [2024-10-29 11:02:26.739748] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:21.493 [2024-10-29 11:02:26.962008] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:13:21.753 114.20 IOPS, 342.60 MiB/s [2024-10-29T11:02:27.251Z] [2024-10-29 11:02:27.074191] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.753 "name": "raid_bdev1", 00:13:21.753 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:21.753 "strip_size_kb": 0, 00:13:21.753 "state": "online", 00:13:21.753 "raid_level": "raid1", 00:13:21.753 "superblock": false, 00:13:21.753 "num_base_bdevs": 4, 00:13:21.753 "num_base_bdevs_discovered": 3, 00:13:21.753 "num_base_bdevs_operational": 3, 00:13:21.753 "process": { 00:13:21.753 "type": "rebuild", 00:13:21.753 "target": "spare", 00:13:21.753 "progress": { 00:13:21.753 "blocks": 34816, 00:13:21.753 "percent": 53 00:13:21.753 } 00:13:21.753 }, 00:13:21.753 "base_bdevs_list": [ 00:13:21.753 { 00:13:21.753 "name": "spare", 00:13:21.753 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:21.753 "is_configured": true, 00:13:21.753 "data_offset": 0, 00:13:21.753 "data_size": 65536 00:13:21.753 }, 00:13:21.753 { 00:13:21.753 "name": null, 00:13:21.753 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.753 "is_configured": false, 00:13:21.753 "data_offset": 0, 00:13:21.753 "data_size": 65536 00:13:21.753 }, 00:13:21.753 { 00:13:21.753 "name": "BaseBdev3", 00:13:21.753 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:21.753 "is_configured": true, 00:13:21.753 "data_offset": 0, 00:13:21.753 "data_size": 65536 00:13:21.753 }, 00:13:21.753 { 00:13:21.753 "name": "BaseBdev4", 00:13:21.753 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:21.753 "is_configured": true, 00:13:21.753 "data_offset": 0, 00:13:21.753 "data_size": 65536 00:13:21.753 } 00:13:21.753 ] 00:13:21.753 }' 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:21.753 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:22.013 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:22.013 11:02:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:22.273 [2024-10-29 11:02:27.737021] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:13:22.844 100.67 IOPS, 302.00 MiB/s [2024-10-29T11:02:28.342Z] [2024-10-29 11:02:28.175528] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.844 "name": "raid_bdev1", 00:13:22.844 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:22.844 "strip_size_kb": 0, 00:13:22.844 "state": "online", 00:13:22.844 "raid_level": "raid1", 00:13:22.844 "superblock": false, 00:13:22.844 "num_base_bdevs": 4, 00:13:22.844 "num_base_bdevs_discovered": 3, 00:13:22.844 "num_base_bdevs_operational": 3, 00:13:22.844 "process": { 00:13:22.844 "type": "rebuild", 00:13:22.844 "target": "spare", 00:13:22.844 "progress": { 00:13:22.844 "blocks": 51200, 00:13:22.844 "percent": 78 00:13:22.844 } 00:13:22.844 }, 00:13:22.844 "base_bdevs_list": [ 00:13:22.844 { 00:13:22.844 "name": "spare", 00:13:22.844 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:22.844 "is_configured": true, 00:13:22.844 "data_offset": 0, 00:13:22.844 "data_size": 65536 00:13:22.844 }, 00:13:22.844 { 00:13:22.844 "name": null, 00:13:22.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:22.844 "is_configured": false, 00:13:22.844 "data_offset": 0, 00:13:22.844 "data_size": 65536 00:13:22.844 }, 00:13:22.844 { 00:13:22.844 "name": "BaseBdev3", 00:13:22.844 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:22.844 "is_configured": true, 00:13:22.844 "data_offset": 0, 00:13:22.844 "data_size": 65536 00:13:22.844 }, 00:13:22.844 { 00:13:22.844 "name": "BaseBdev4", 00:13:22.844 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:22.844 "is_configured": true, 00:13:22.844 "data_offset": 0, 00:13:22.844 "data_size": 65536 00:13:22.844 } 00:13:22.844 ] 00:13:22.844 }' 00:13:22.844 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.104 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:23.104 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.104 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:23.104 11:02:28 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:23.674 [2024-10-29 11:02:28.922937] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:23.674 [2024-10-29 11:02:29.027764] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:23.674 [2024-10-29 11:02:29.029953] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:23.934 93.43 IOPS, 280.29 MiB/s [2024-10-29T11:02:29.432Z] 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.934 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:24.195 "name": "raid_bdev1", 00:13:24.195 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:24.195 "strip_size_kb": 0, 00:13:24.195 "state": "online", 00:13:24.195 "raid_level": "raid1", 00:13:24.195 "superblock": false, 00:13:24.195 "num_base_bdevs": 4, 00:13:24.195 "num_base_bdevs_discovered": 3, 00:13:24.195 "num_base_bdevs_operational": 3, 00:13:24.195 "base_bdevs_list": [ 00:13:24.195 { 00:13:24.195 "name": "spare", 00:13:24.195 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": null, 00:13:24.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.195 "is_configured": false, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": "BaseBdev3", 00:13:24.195 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": "BaseBdev4", 00:13:24.195 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 } 00:13:24.195 ] 00:13:24.195 }' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:24.195 "name": "raid_bdev1", 00:13:24.195 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:24.195 "strip_size_kb": 0, 00:13:24.195 "state": "online", 00:13:24.195 "raid_level": "raid1", 00:13:24.195 "superblock": false, 00:13:24.195 "num_base_bdevs": 4, 00:13:24.195 "num_base_bdevs_discovered": 3, 00:13:24.195 "num_base_bdevs_operational": 3, 00:13:24.195 "base_bdevs_list": [ 00:13:24.195 { 00:13:24.195 "name": "spare", 00:13:24.195 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": null, 00:13:24.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.195 "is_configured": false, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": "BaseBdev3", 00:13:24.195 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 }, 00:13:24.195 { 00:13:24.195 "name": "BaseBdev4", 00:13:24.195 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:24.195 "is_configured": true, 00:13:24.195 "data_offset": 0, 00:13:24.195 "data_size": 65536 00:13:24.195 } 00:13:24.195 ] 00:13:24.195 }' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.195 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.455 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.455 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.455 "name": "raid_bdev1", 00:13:24.455 "uuid": "6f361573-c244-4e33-b129-e014b3e5bfb5", 00:13:24.455 "strip_size_kb": 0, 00:13:24.455 "state": "online", 00:13:24.455 "raid_level": "raid1", 00:13:24.455 "superblock": false, 00:13:24.455 "num_base_bdevs": 4, 00:13:24.455 "num_base_bdevs_discovered": 3, 00:13:24.455 "num_base_bdevs_operational": 3, 00:13:24.455 "base_bdevs_list": [ 00:13:24.455 { 00:13:24.455 "name": "spare", 00:13:24.455 "uuid": "a1e31aa0-f970-5c50-8e91-146442a2f3aa", 00:13:24.455 "is_configured": true, 00:13:24.455 "data_offset": 0, 00:13:24.455 "data_size": 65536 00:13:24.455 }, 00:13:24.455 { 00:13:24.455 "name": null, 00:13:24.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.455 "is_configured": false, 00:13:24.455 "data_offset": 0, 00:13:24.455 "data_size": 65536 00:13:24.455 }, 00:13:24.455 { 00:13:24.455 "name": "BaseBdev3", 00:13:24.455 "uuid": "da506571-7045-5edf-9dd4-27c42a867686", 00:13:24.455 "is_configured": true, 00:13:24.455 "data_offset": 0, 00:13:24.455 "data_size": 65536 00:13:24.455 }, 00:13:24.455 { 00:13:24.455 "name": "BaseBdev4", 00:13:24.455 "uuid": "eaa179da-b81b-5875-a363-18f832f836ee", 00:13:24.455 "is_configured": true, 00:13:24.455 "data_offset": 0, 00:13:24.455 "data_size": 65536 00:13:24.455 } 00:13:24.455 ] 00:13:24.455 }' 00:13:24.455 11:02:29 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.455 11:02:29 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.714 85.25 IOPS, 255.75 MiB/s [2024-10-29T11:02:30.212Z] 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.714 [2024-10-29 11:02:30.074504] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:24.714 [2024-10-29 11:02:30.074539] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:24.714 00:13:24.714 Latency(us) 00:13:24.714 [2024-10-29T11:02:30.212Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:24.714 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:24.714 raid_bdev1 : 8.08 84.61 253.84 0.00 0.00 16016.21 304.07 116304.94 00:13:24.714 [2024-10-29T11:02:30.212Z] =================================================================================================================== 00:13:24.714 [2024-10-29T11:02:30.212Z] Total : 84.61 253.84 0.00 0.00 16016.21 304.07 116304.94 00:13:24.714 { 00:13:24.714 "results": [ 00:13:24.714 { 00:13:24.714 "job": "raid_bdev1", 00:13:24.714 "core_mask": "0x1", 00:13:24.714 "workload": "randrw", 00:13:24.714 "percentage": 50, 00:13:24.714 "status": "finished", 00:13:24.714 "queue_depth": 2, 00:13:24.714 "io_size": 3145728, 00:13:24.714 "runtime": 8.08373, 00:13:24.714 "iops": 84.61440448901683, 00:13:24.714 "mibps": 253.8432134670505, 00:13:24.714 "io_failed": 0, 00:13:24.714 "io_timeout": 0, 00:13:24.714 "avg_latency_us": 16016.210914476876, 00:13:24.714 "min_latency_us": 304.0698689956332, 00:13:24.714 "max_latency_us": 116304.93624454149 00:13:24.714 } 00:13:24.714 ], 00:13:24.714 "core_count": 1 00:13:24.714 } 00:13:24.714 [2024-10-29 11:02:30.108484] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:24.714 [2024-10-29 11:02:30.108530] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:24.714 [2024-10-29 11:02:30.108644] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:24.714 [2024-10-29 11:02:30.108658] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:24.714 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:24.715 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:24.715 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:24.715 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:24.975 /dev/nbd0 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # local i 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # break 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:24.975 1+0 records in 00:13:24.975 1+0 records out 00:13:24.975 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00031735 s, 12.9 MB/s 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # size=4096 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # return 0 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:24.975 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:24.976 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:25.236 /dev/nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # local i 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # break 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:25.236 1+0 records in 00:13:25.236 1+0 records out 00:13:25.236 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000346928 s, 11.8 MB/s 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # size=4096 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # return 0 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:25.236 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:25.497 11:02:30 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:25.757 /dev/nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # local i 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # break 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:25.757 1+0 records in 00:13:25.757 1+0 records out 00:13:25.757 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000306705 s, 13.4 MB/s 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # size=4096 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # return 0 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:25.757 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:26.017 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 89468 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@952 -- # '[' -z 89468 ']' 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # kill -0 89468 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@957 -- # uname 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 89468 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@970 -- # echo 'killing process with pid 89468' 00:13:26.278 killing process with pid 89468 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@971 -- # kill 89468 00:13:26.278 Received shutdown signal, test time was about 9.644428 seconds 00:13:26.278 00:13:26.278 Latency(us) 00:13:26.278 [2024-10-29T11:02:31.776Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:26.278 [2024-10-29T11:02:31.776Z] =================================================================================================================== 00:13:26.278 [2024-10-29T11:02:31.776Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:26.278 [2024-10-29 11:02:31.663055] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:26.278 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@976 -- # wait 89468 00:13:26.278 [2024-10-29 11:02:31.708333] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:26.538 00:13:26.538 real 0m11.593s 00:13:26.538 user 0m14.910s 00:13:26.538 sys 0m1.756s 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1128 -- # xtrace_disable 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.538 ************************************ 00:13:26.538 END TEST raid_rebuild_test_io 00:13:26.538 ************************************ 00:13:26.538 11:02:31 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:13:26.538 11:02:31 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:13:26.538 11:02:31 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:13:26.538 11:02:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:26.538 ************************************ 00:13:26.538 START TEST raid_rebuild_test_sb_io 00:13:26.538 ************************************ 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 4 true true true 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:26.538 11:02:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89861 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89861 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@833 -- # '[' -z 89861 ']' 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # local max_retries=100 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:26.538 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # xtrace_disable 00:13:26.538 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.798 [2024-10-29 11:02:32.084735] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:13:26.798 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:26.798 Zero copy mechanism will not be used. 00:13:26.798 [2024-10-29 11:02:32.084964] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89861 ] 00:13:26.798 [2024-10-29 11:02:32.255156] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:26.798 [2024-10-29 11:02:32.279815] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:27.058 [2024-10-29 11:02:32.323051] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:27.058 [2024-10-29 11:02:32.323186] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@866 -- # return 0 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.628 BaseBdev1_malloc 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.628 [2024-10-29 11:02:32.922768] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:27.628 [2024-10-29 11:02:32.922882] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.628 [2024-10-29 11:02:32.922941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:27.628 [2024-10-29 11:02:32.923003] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.628 [2024-10-29 11:02:32.925199] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.628 [2024-10-29 11:02:32.925281] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:27.628 BaseBdev1 00:13:27.628 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 BaseBdev2_malloc 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 [2024-10-29 11:02:32.951809] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:27.629 [2024-10-29 11:02:32.951910] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.629 [2024-10-29 11:02:32.951964] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:27.629 [2024-10-29 11:02:32.952000] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.629 [2024-10-29 11:02:32.954081] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.629 [2024-10-29 11:02:32.954160] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:27.629 BaseBdev2 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 BaseBdev3_malloc 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 [2024-10-29 11:02:32.980460] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:27.629 [2024-10-29 11:02:32.980514] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.629 [2024-10-29 11:02:32.980541] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:27.629 [2024-10-29 11:02:32.980553] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.629 [2024-10-29 11:02:32.982574] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.629 [2024-10-29 11:02:32.982610] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:27.629 BaseBdev3 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 BaseBdev4_malloc 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 [2024-10-29 11:02:33.024663] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:27.629 [2024-10-29 11:02:33.024758] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.629 [2024-10-29 11:02:33.024807] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:27.629 [2024-10-29 11:02:33.024831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.629 [2024-10-29 11:02:33.029125] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.629 [2024-10-29 11:02:33.029287] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:27.629 BaseBdev4 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 spare_malloc 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 spare_delay 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 [2024-10-29 11:02:33.066782] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:27.629 [2024-10-29 11:02:33.066835] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.629 [2024-10-29 11:02:33.066875] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:27.629 [2024-10-29 11:02:33.066884] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.629 [2024-10-29 11:02:33.068962] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.629 [2024-10-29 11:02:33.069008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:27.629 spare 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 [2024-10-29 11:02:33.078807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:27.629 [2024-10-29 11:02:33.080661] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:27.629 [2024-10-29 11:02:33.080734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:27.629 [2024-10-29 11:02:33.080778] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:27.629 [2024-10-29 11:02:33.080952] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:13:27.629 [2024-10-29 11:02:33.080964] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:27.629 [2024-10-29 11:02:33.081230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:13:27.629 [2024-10-29 11:02:33.081393] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:13:27.629 [2024-10-29 11:02:33.081410] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:13:27.629 [2024-10-29 11:02:33.081558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.629 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.889 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.889 "name": "raid_bdev1", 00:13:27.889 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:27.889 "strip_size_kb": 0, 00:13:27.889 "state": "online", 00:13:27.889 "raid_level": "raid1", 00:13:27.889 "superblock": true, 00:13:27.890 "num_base_bdevs": 4, 00:13:27.890 "num_base_bdevs_discovered": 4, 00:13:27.890 "num_base_bdevs_operational": 4, 00:13:27.890 "base_bdevs_list": [ 00:13:27.890 { 00:13:27.890 "name": "BaseBdev1", 00:13:27.890 "uuid": "ca80074e-9c27-5a7a-af70-2c22c5f2de39", 00:13:27.890 "is_configured": true, 00:13:27.890 "data_offset": 2048, 00:13:27.890 "data_size": 63488 00:13:27.890 }, 00:13:27.890 { 00:13:27.890 "name": "BaseBdev2", 00:13:27.890 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:27.890 "is_configured": true, 00:13:27.890 "data_offset": 2048, 00:13:27.890 "data_size": 63488 00:13:27.890 }, 00:13:27.890 { 00:13:27.890 "name": "BaseBdev3", 00:13:27.890 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:27.890 "is_configured": true, 00:13:27.890 "data_offset": 2048, 00:13:27.890 "data_size": 63488 00:13:27.890 }, 00:13:27.890 { 00:13:27.890 "name": "BaseBdev4", 00:13:27.890 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:27.890 "is_configured": true, 00:13:27.890 "data_offset": 2048, 00:13:27.890 "data_size": 63488 00:13:27.890 } 00:13:27.890 ] 00:13:27.890 }' 00:13:27.890 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.890 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 [2024-10-29 11:02:33.482456] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 [2024-10-29 11:02:33.569966] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.150 "name": "raid_bdev1", 00:13:28.150 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:28.150 "strip_size_kb": 0, 00:13:28.150 "state": "online", 00:13:28.150 "raid_level": "raid1", 00:13:28.150 "superblock": true, 00:13:28.150 "num_base_bdevs": 4, 00:13:28.150 "num_base_bdevs_discovered": 3, 00:13:28.150 "num_base_bdevs_operational": 3, 00:13:28.150 "base_bdevs_list": [ 00:13:28.150 { 00:13:28.150 "name": null, 00:13:28.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.150 "is_configured": false, 00:13:28.150 "data_offset": 0, 00:13:28.150 "data_size": 63488 00:13:28.150 }, 00:13:28.150 { 00:13:28.150 "name": "BaseBdev2", 00:13:28.150 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:28.150 "is_configured": true, 00:13:28.150 "data_offset": 2048, 00:13:28.150 "data_size": 63488 00:13:28.150 }, 00:13:28.150 { 00:13:28.150 "name": "BaseBdev3", 00:13:28.150 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:28.150 "is_configured": true, 00:13:28.150 "data_offset": 2048, 00:13:28.150 "data_size": 63488 00:13:28.150 }, 00:13:28.150 { 00:13:28.150 "name": "BaseBdev4", 00:13:28.150 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:28.150 "is_configured": true, 00:13:28.150 "data_offset": 2048, 00:13:28.150 "data_size": 63488 00:13:28.150 } 00:13:28.150 ] 00:13:28.150 }' 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.150 11:02:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.150 [2024-10-29 11:02:33.647916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:28.410 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:28.410 Zero copy mechanism will not be used. 00:13:28.410 Running I/O for 60 seconds... 00:13:28.670 11:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:28.670 11:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.670 11:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:28.670 [2024-10-29 11:02:34.024982] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:28.670 11:02:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.670 11:02:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:28.670 [2024-10-29 11:02:34.067037] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:13:28.670 [2024-10-29 11:02:34.069064] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:28.930 [2024-10-29 11:02:34.210176] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:28.930 [2024-10-29 11:02:34.211424] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:28.930 [2024-10-29 11:02:34.427075] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:28.930 [2024-10-29 11:02:34.427806] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:29.449 185.00 IOPS, 555.00 MiB/s [2024-10-29T11:02:34.947Z] [2024-10-29 11:02:34.785757] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:29.710 [2024-10-29 11:02:35.008917] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:29.710 [2024-10-29 11:02:35.009585] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:29.710 "name": "raid_bdev1", 00:13:29.710 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:29.710 "strip_size_kb": 0, 00:13:29.710 "state": "online", 00:13:29.710 "raid_level": "raid1", 00:13:29.710 "superblock": true, 00:13:29.710 "num_base_bdevs": 4, 00:13:29.710 "num_base_bdevs_discovered": 4, 00:13:29.710 "num_base_bdevs_operational": 4, 00:13:29.710 "process": { 00:13:29.710 "type": "rebuild", 00:13:29.710 "target": "spare", 00:13:29.710 "progress": { 00:13:29.710 "blocks": 10240, 00:13:29.710 "percent": 16 00:13:29.710 } 00:13:29.710 }, 00:13:29.710 "base_bdevs_list": [ 00:13:29.710 { 00:13:29.710 "name": "spare", 00:13:29.710 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:29.710 "is_configured": true, 00:13:29.710 "data_offset": 2048, 00:13:29.710 "data_size": 63488 00:13:29.710 }, 00:13:29.710 { 00:13:29.710 "name": "BaseBdev2", 00:13:29.710 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:29.710 "is_configured": true, 00:13:29.710 "data_offset": 2048, 00:13:29.710 "data_size": 63488 00:13:29.710 }, 00:13:29.710 { 00:13:29.710 "name": "BaseBdev3", 00:13:29.710 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:29.710 "is_configured": true, 00:13:29.710 "data_offset": 2048, 00:13:29.710 "data_size": 63488 00:13:29.710 }, 00:13:29.710 { 00:13:29.710 "name": "BaseBdev4", 00:13:29.710 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:29.710 "is_configured": true, 00:13:29.710 "data_offset": 2048, 00:13:29.710 "data_size": 63488 00:13:29.710 } 00:13:29.710 ] 00:13:29.710 }' 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.710 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.710 [2024-10-29 11:02:35.169387] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:29.970 [2024-10-29 11:02:35.329090] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:29.970 [2024-10-29 11:02:35.332258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:29.970 [2024-10-29 11:02:35.332297] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:29.970 [2024-10-29 11:02:35.332312] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:29.970 [2024-10-29 11:02:35.349721] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.970 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:29.970 "name": "raid_bdev1", 00:13:29.970 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:29.970 "strip_size_kb": 0, 00:13:29.970 "state": "online", 00:13:29.970 "raid_level": "raid1", 00:13:29.970 "superblock": true, 00:13:29.970 "num_base_bdevs": 4, 00:13:29.970 "num_base_bdevs_discovered": 3, 00:13:29.970 "num_base_bdevs_operational": 3, 00:13:29.970 "base_bdevs_list": [ 00:13:29.970 { 00:13:29.970 "name": null, 00:13:29.970 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.970 "is_configured": false, 00:13:29.970 "data_offset": 0, 00:13:29.970 "data_size": 63488 00:13:29.970 }, 00:13:29.970 { 00:13:29.970 "name": "BaseBdev2", 00:13:29.970 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:29.970 "is_configured": true, 00:13:29.970 "data_offset": 2048, 00:13:29.970 "data_size": 63488 00:13:29.970 }, 00:13:29.970 { 00:13:29.970 "name": "BaseBdev3", 00:13:29.970 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:29.970 "is_configured": true, 00:13:29.970 "data_offset": 2048, 00:13:29.970 "data_size": 63488 00:13:29.970 }, 00:13:29.970 { 00:13:29.970 "name": "BaseBdev4", 00:13:29.971 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:29.971 "is_configured": true, 00:13:29.971 "data_offset": 2048, 00:13:29.971 "data_size": 63488 00:13:29.971 } 00:13:29.971 ] 00:13:29.971 }' 00:13:29.971 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:29.971 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.490 163.00 IOPS, 489.00 MiB/s [2024-10-29T11:02:35.988Z] 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:30.490 "name": "raid_bdev1", 00:13:30.490 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:30.490 "strip_size_kb": 0, 00:13:30.490 "state": "online", 00:13:30.490 "raid_level": "raid1", 00:13:30.490 "superblock": true, 00:13:30.490 "num_base_bdevs": 4, 00:13:30.490 "num_base_bdevs_discovered": 3, 00:13:30.490 "num_base_bdevs_operational": 3, 00:13:30.490 "base_bdevs_list": [ 00:13:30.490 { 00:13:30.490 "name": null, 00:13:30.490 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.490 "is_configured": false, 00:13:30.490 "data_offset": 0, 00:13:30.490 "data_size": 63488 00:13:30.490 }, 00:13:30.490 { 00:13:30.490 "name": "BaseBdev2", 00:13:30.490 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:30.490 "is_configured": true, 00:13:30.490 "data_offset": 2048, 00:13:30.490 "data_size": 63488 00:13:30.490 }, 00:13:30.490 { 00:13:30.490 "name": "BaseBdev3", 00:13:30.490 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:30.490 "is_configured": true, 00:13:30.490 "data_offset": 2048, 00:13:30.490 "data_size": 63488 00:13:30.490 }, 00:13:30.490 { 00:13:30.490 "name": "BaseBdev4", 00:13:30.490 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:30.490 "is_configured": true, 00:13:30.490 "data_offset": 2048, 00:13:30.490 "data_size": 63488 00:13:30.490 } 00:13:30.490 ] 00:13:30.490 }' 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.490 [2024-10-29 11:02:35.961548] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.490 11:02:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:30.750 [2024-10-29 11:02:36.002159] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:13:30.750 [2024-10-29 11:02:36.004275] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:30.750 [2024-10-29 11:02:36.118776] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:30.750 [2024-10-29 11:02:36.119445] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:30.750 [2024-10-29 11:02:36.248141] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:30.750 [2024-10-29 11:02:36.248955] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:31.319 [2024-10-29 11:02:36.613100] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:31.319 165.00 IOPS, 495.00 MiB/s [2024-10-29T11:02:36.817Z] [2024-10-29 11:02:36.739846] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:31.319 [2024-10-29 11:02:36.740556] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.579 11:02:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.579 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.579 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:31.579 "name": "raid_bdev1", 00:13:31.579 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:31.579 "strip_size_kb": 0, 00:13:31.579 "state": "online", 00:13:31.579 "raid_level": "raid1", 00:13:31.579 "superblock": true, 00:13:31.579 "num_base_bdevs": 4, 00:13:31.579 "num_base_bdevs_discovered": 4, 00:13:31.579 "num_base_bdevs_operational": 4, 00:13:31.579 "process": { 00:13:31.579 "type": "rebuild", 00:13:31.579 "target": "spare", 00:13:31.579 "progress": { 00:13:31.579 "blocks": 12288, 00:13:31.579 "percent": 19 00:13:31.579 } 00:13:31.579 }, 00:13:31.579 "base_bdevs_list": [ 00:13:31.579 { 00:13:31.579 "name": "spare", 00:13:31.579 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:31.579 "is_configured": true, 00:13:31.579 "data_offset": 2048, 00:13:31.579 "data_size": 63488 00:13:31.579 }, 00:13:31.580 { 00:13:31.580 "name": "BaseBdev2", 00:13:31.580 "uuid": "34a723cd-ed88-5552-9799-2b258063e634", 00:13:31.580 "is_configured": true, 00:13:31.580 "data_offset": 2048, 00:13:31.580 "data_size": 63488 00:13:31.580 }, 00:13:31.580 { 00:13:31.580 "name": "BaseBdev3", 00:13:31.580 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:31.580 "is_configured": true, 00:13:31.580 "data_offset": 2048, 00:13:31.580 "data_size": 63488 00:13:31.580 }, 00:13:31.580 { 00:13:31.580 "name": "BaseBdev4", 00:13:31.580 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:31.580 "is_configured": true, 00:13:31.580 "data_offset": 2048, 00:13:31.580 "data_size": 63488 00:13:31.580 } 00:13:31.580 ] 00:13:31.580 }' 00:13:31.580 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:31.839 [2024-10-29 11:02:37.085089] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:31.839 [2024-10-29 11:02:37.085547] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:31.839 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.839 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.839 [2024-10-29 11:02:37.127182] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:31.839 [2024-10-29 11:02:37.296649] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:31.839 [2024-10-29 11:02:37.297049] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:32.100 [2024-10-29 11:02:37.504607] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006080 00:13:32.100 [2024-10-29 11:02:37.504714] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:32.100 [2024-10-29 11:02:37.524931] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:32.100 "name": "raid_bdev1", 00:13:32.100 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:32.100 "strip_size_kb": 0, 00:13:32.100 "state": "online", 00:13:32.100 "raid_level": "raid1", 00:13:32.100 "superblock": true, 00:13:32.100 "num_base_bdevs": 4, 00:13:32.100 "num_base_bdevs_discovered": 3, 00:13:32.100 "num_base_bdevs_operational": 3, 00:13:32.100 "process": { 00:13:32.100 "type": "rebuild", 00:13:32.100 "target": "spare", 00:13:32.100 "progress": { 00:13:32.100 "blocks": 16384, 00:13:32.100 "percent": 25 00:13:32.100 } 00:13:32.100 }, 00:13:32.100 "base_bdevs_list": [ 00:13:32.100 { 00:13:32.100 "name": "spare", 00:13:32.100 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:32.100 "is_configured": true, 00:13:32.100 "data_offset": 2048, 00:13:32.100 "data_size": 63488 00:13:32.100 }, 00:13:32.100 { 00:13:32.100 "name": null, 00:13:32.100 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.100 "is_configured": false, 00:13:32.100 "data_offset": 0, 00:13:32.100 "data_size": 63488 00:13:32.100 }, 00:13:32.100 { 00:13:32.100 "name": "BaseBdev3", 00:13:32.100 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:32.100 "is_configured": true, 00:13:32.100 "data_offset": 2048, 00:13:32.100 "data_size": 63488 00:13:32.100 }, 00:13:32.100 { 00:13:32.100 "name": "BaseBdev4", 00:13:32.100 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:32.100 "is_configured": true, 00:13:32.100 "data_offset": 2048, 00:13:32.100 "data_size": 63488 00:13:32.100 } 00:13:32.100 ] 00:13:32.100 }' 00:13:32.100 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:32.361 135.00 IOPS, 405.00 MiB/s [2024-10-29T11:02:37.859Z] 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=402 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:32.361 "name": "raid_bdev1", 00:13:32.361 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:32.361 "strip_size_kb": 0, 00:13:32.361 "state": "online", 00:13:32.361 "raid_level": "raid1", 00:13:32.361 "superblock": true, 00:13:32.361 "num_base_bdevs": 4, 00:13:32.361 "num_base_bdevs_discovered": 3, 00:13:32.361 "num_base_bdevs_operational": 3, 00:13:32.361 "process": { 00:13:32.361 "type": "rebuild", 00:13:32.361 "target": "spare", 00:13:32.361 "progress": { 00:13:32.361 "blocks": 18432, 00:13:32.361 "percent": 29 00:13:32.361 } 00:13:32.361 }, 00:13:32.361 "base_bdevs_list": [ 00:13:32.361 { 00:13:32.361 "name": "spare", 00:13:32.361 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:32.361 "is_configured": true, 00:13:32.361 "data_offset": 2048, 00:13:32.361 "data_size": 63488 00:13:32.361 }, 00:13:32.361 { 00:13:32.361 "name": null, 00:13:32.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.361 "is_configured": false, 00:13:32.361 "data_offset": 0, 00:13:32.361 "data_size": 63488 00:13:32.361 }, 00:13:32.361 { 00:13:32.361 "name": "BaseBdev3", 00:13:32.361 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:32.361 "is_configured": true, 00:13:32.361 "data_offset": 2048, 00:13:32.361 "data_size": 63488 00:13:32.361 }, 00:13:32.361 { 00:13:32.361 "name": "BaseBdev4", 00:13:32.361 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:32.361 "is_configured": true, 00:13:32.361 "data_offset": 2048, 00:13:32.361 "data_size": 63488 00:13:32.361 } 00:13:32.361 ] 00:13:32.361 }' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:32.361 11:02:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:32.361 [2024-10-29 11:02:37.766336] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:32.621 [2024-10-29 11:02:37.892942] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:32.881 [2024-10-29 11:02:38.128119] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:33.450 124.20 IOPS, 372.60 MiB/s [2024-10-29T11:02:38.948Z] 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:33.450 "name": "raid_bdev1", 00:13:33.450 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:33.450 "strip_size_kb": 0, 00:13:33.450 "state": "online", 00:13:33.450 "raid_level": "raid1", 00:13:33.450 "superblock": true, 00:13:33.450 "num_base_bdevs": 4, 00:13:33.450 "num_base_bdevs_discovered": 3, 00:13:33.450 "num_base_bdevs_operational": 3, 00:13:33.450 "process": { 00:13:33.450 "type": "rebuild", 00:13:33.450 "target": "spare", 00:13:33.450 "progress": { 00:13:33.450 "blocks": 36864, 00:13:33.450 "percent": 58 00:13:33.450 } 00:13:33.450 }, 00:13:33.450 "base_bdevs_list": [ 00:13:33.450 { 00:13:33.450 "name": "spare", 00:13:33.450 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:33.450 "is_configured": true, 00:13:33.450 "data_offset": 2048, 00:13:33.450 "data_size": 63488 00:13:33.450 }, 00:13:33.450 { 00:13:33.450 "name": null, 00:13:33.450 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.450 "is_configured": false, 00:13:33.450 "data_offset": 0, 00:13:33.450 "data_size": 63488 00:13:33.450 }, 00:13:33.450 { 00:13:33.450 "name": "BaseBdev3", 00:13:33.450 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:33.450 "is_configured": true, 00:13:33.450 "data_offset": 2048, 00:13:33.450 "data_size": 63488 00:13:33.450 }, 00:13:33.450 { 00:13:33.450 "name": "BaseBdev4", 00:13:33.450 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:33.450 "is_configured": true, 00:13:33.450 "data_offset": 2048, 00:13:33.450 "data_size": 63488 00:13:33.450 } 00:13:33.450 ] 00:13:33.450 }' 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:33.450 [2024-10-29 11:02:38.898512] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:33.450 11:02:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:33.710 [2024-10-29 11:02:39.109857] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:13:33.970 [2024-10-29 11:02:39.222936] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:34.231 [2024-10-29 11:02:39.536052] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:13:34.232 [2024-10-29 11:02:39.642505] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:34.496 108.67 IOPS, 326.00 MiB/s [2024-10-29T11:02:39.994Z] 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:34.496 "name": "raid_bdev1", 00:13:34.496 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:34.496 "strip_size_kb": 0, 00:13:34.496 "state": "online", 00:13:34.496 "raid_level": "raid1", 00:13:34.496 "superblock": true, 00:13:34.496 "num_base_bdevs": 4, 00:13:34.496 "num_base_bdevs_discovered": 3, 00:13:34.496 "num_base_bdevs_operational": 3, 00:13:34.496 "process": { 00:13:34.496 "type": "rebuild", 00:13:34.496 "target": "spare", 00:13:34.496 "progress": { 00:13:34.496 "blocks": 57344, 00:13:34.496 "percent": 90 00:13:34.496 } 00:13:34.496 }, 00:13:34.496 "base_bdevs_list": [ 00:13:34.496 { 00:13:34.496 "name": "spare", 00:13:34.496 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:34.496 "is_configured": true, 00:13:34.496 "data_offset": 2048, 00:13:34.496 "data_size": 63488 00:13:34.496 }, 00:13:34.496 { 00:13:34.496 "name": null, 00:13:34.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.496 "is_configured": false, 00:13:34.496 "data_offset": 0, 00:13:34.496 "data_size": 63488 00:13:34.496 }, 00:13:34.496 { 00:13:34.496 "name": "BaseBdev3", 00:13:34.496 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:34.496 "is_configured": true, 00:13:34.496 "data_offset": 2048, 00:13:34.496 "data_size": 63488 00:13:34.496 }, 00:13:34.496 { 00:13:34.496 "name": "BaseBdev4", 00:13:34.496 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:34.496 "is_configured": true, 00:13:34.496 "data_offset": 2048, 00:13:34.496 "data_size": 63488 00:13:34.496 } 00:13:34.496 ] 00:13:34.496 }' 00:13:34.496 11:02:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:34.496 [2024-10-29 11:02:39.964270] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:34.496 [2024-10-29 11:02:39.964514] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:34.756 11:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:34.756 11:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:34.756 11:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:34.756 11:02:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:34.756 [2024-10-29 11:02:40.201984] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:35.046 [2024-10-29 11:02:40.306811] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:35.046 [2024-10-29 11:02:40.309080] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:35.635 98.43 IOPS, 295.29 MiB/s [2024-10-29T11:02:41.133Z] 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.635 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.635 "name": "raid_bdev1", 00:13:35.635 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:35.635 "strip_size_kb": 0, 00:13:35.635 "state": "online", 00:13:35.635 "raid_level": "raid1", 00:13:35.635 "superblock": true, 00:13:35.635 "num_base_bdevs": 4, 00:13:35.635 "num_base_bdevs_discovered": 3, 00:13:35.635 "num_base_bdevs_operational": 3, 00:13:35.635 "base_bdevs_list": [ 00:13:35.635 { 00:13:35.635 "name": "spare", 00:13:35.635 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:35.635 "is_configured": true, 00:13:35.635 "data_offset": 2048, 00:13:35.635 "data_size": 63488 00:13:35.635 }, 00:13:35.635 { 00:13:35.635 "name": null, 00:13:35.635 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.635 "is_configured": false, 00:13:35.635 "data_offset": 0, 00:13:35.636 "data_size": 63488 00:13:35.636 }, 00:13:35.636 { 00:13:35.636 "name": "BaseBdev3", 00:13:35.636 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:35.636 "is_configured": true, 00:13:35.636 "data_offset": 2048, 00:13:35.636 "data_size": 63488 00:13:35.636 }, 00:13:35.636 { 00:13:35.636 "name": "BaseBdev4", 00:13:35.636 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:35.636 "is_configured": true, 00:13:35.636 "data_offset": 2048, 00:13:35.636 "data_size": 63488 00:13:35.636 } 00:13:35.636 ] 00:13:35.636 }' 00:13:35.636 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.896 "name": "raid_bdev1", 00:13:35.896 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:35.896 "strip_size_kb": 0, 00:13:35.896 "state": "online", 00:13:35.896 "raid_level": "raid1", 00:13:35.896 "superblock": true, 00:13:35.896 "num_base_bdevs": 4, 00:13:35.896 "num_base_bdevs_discovered": 3, 00:13:35.896 "num_base_bdevs_operational": 3, 00:13:35.896 "base_bdevs_list": [ 00:13:35.896 { 00:13:35.896 "name": "spare", 00:13:35.896 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:35.896 "is_configured": true, 00:13:35.896 "data_offset": 2048, 00:13:35.896 "data_size": 63488 00:13:35.896 }, 00:13:35.896 { 00:13:35.896 "name": null, 00:13:35.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.896 "is_configured": false, 00:13:35.896 "data_offset": 0, 00:13:35.896 "data_size": 63488 00:13:35.896 }, 00:13:35.896 { 00:13:35.896 "name": "BaseBdev3", 00:13:35.896 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:35.896 "is_configured": true, 00:13:35.896 "data_offset": 2048, 00:13:35.896 "data_size": 63488 00:13:35.896 }, 00:13:35.896 { 00:13:35.896 "name": "BaseBdev4", 00:13:35.896 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:35.896 "is_configured": true, 00:13:35.896 "data_offset": 2048, 00:13:35.896 "data_size": 63488 00:13:35.896 } 00:13:35.896 ] 00:13:35.896 }' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.896 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.897 "name": "raid_bdev1", 00:13:35.897 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:35.897 "strip_size_kb": 0, 00:13:35.897 "state": "online", 00:13:35.897 "raid_level": "raid1", 00:13:35.897 "superblock": true, 00:13:35.897 "num_base_bdevs": 4, 00:13:35.897 "num_base_bdevs_discovered": 3, 00:13:35.897 "num_base_bdevs_operational": 3, 00:13:35.897 "base_bdevs_list": [ 00:13:35.897 { 00:13:35.897 "name": "spare", 00:13:35.897 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:35.897 "is_configured": true, 00:13:35.897 "data_offset": 2048, 00:13:35.897 "data_size": 63488 00:13:35.897 }, 00:13:35.897 { 00:13:35.897 "name": null, 00:13:35.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.897 "is_configured": false, 00:13:35.897 "data_offset": 0, 00:13:35.897 "data_size": 63488 00:13:35.897 }, 00:13:35.897 { 00:13:35.897 "name": "BaseBdev3", 00:13:35.897 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:35.897 "is_configured": true, 00:13:35.897 "data_offset": 2048, 00:13:35.897 "data_size": 63488 00:13:35.897 }, 00:13:35.897 { 00:13:35.897 "name": "BaseBdev4", 00:13:35.897 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:35.897 "is_configured": true, 00:13:35.897 "data_offset": 2048, 00:13:35.897 "data_size": 63488 00:13:35.897 } 00:13:35.897 ] 00:13:35.897 }' 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.897 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:36.418 92.12 IOPS, 276.38 MiB/s [2024-10-29T11:02:41.916Z] 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:36.418 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.418 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:36.418 [2024-10-29 11:02:41.730774] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:36.418 [2024-10-29 11:02:41.730819] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:36.418 00:13:36.418 Latency(us) 00:13:36.418 [2024-10-29T11:02:41.917Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:36.419 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:36.419 raid_bdev1 : 8.18 90.75 272.25 0.00 0.00 14529.72 287.97 114473.36 00:13:36.419 [2024-10-29T11:02:41.917Z] =================================================================================================================== 00:13:36.419 [2024-10-29T11:02:41.917Z] Total : 90.75 272.25 0.00 0.00 14529.72 287.97 114473.36 00:13:36.419 [2024-10-29 11:02:41.813651] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:36.419 [2024-10-29 11:02:41.813703] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:36.419 [2024-10-29 11:02:41.813839] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:36.419 [2024-10-29 11:02:41.813863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:13:36.419 { 00:13:36.419 "results": [ 00:13:36.419 { 00:13:36.419 "job": "raid_bdev1", 00:13:36.419 "core_mask": "0x1", 00:13:36.419 "workload": "randrw", 00:13:36.419 "percentage": 50, 00:13:36.419 "status": "finished", 00:13:36.419 "queue_depth": 2, 00:13:36.419 "io_size": 3145728, 00:13:36.419 "runtime": 8.176225, 00:13:36.419 "iops": 90.7509272310877, 00:13:36.419 "mibps": 272.2527816932631, 00:13:36.419 "io_failed": 0, 00:13:36.419 "io_timeout": 0, 00:13:36.419 "avg_latency_us": 14529.71617839193, 00:13:36.419 "min_latency_us": 287.97205240174674, 00:13:36.419 "max_latency_us": 114473.36244541485 00:13:36.419 } 00:13:36.419 ], 00:13:36.419 "core_count": 1 00:13:36.419 } 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.419 11:02:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:36.680 /dev/nbd0 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # local i 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # break 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:36.680 1+0 records in 00:13:36.680 1+0 records out 00:13:36.680 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000469731 s, 8.7 MB/s 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # size=4096 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # return 0 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.680 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:36.941 /dev/nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # local i 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # break 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:36.941 1+0 records in 00:13:36.941 1+0 records out 00:13:36.941 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000329431 s, 12.4 MB/s 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # size=4096 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # return 0 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:36.941 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:37.202 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:37.462 /dev/nbd1 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # local i 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # break 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:13:37.462 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:37.462 1+0 records in 00:13:37.462 1+0 records out 00:13:37.462 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000442223 s, 9.3 MB/s 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # size=4096 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # return 0 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:37.463 11:02:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:37.723 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:37.984 [2024-10-29 11:02:43.328170] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:37.984 [2024-10-29 11:02:43.328236] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:37.984 [2024-10-29 11:02:43.328274] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:13:37.984 [2024-10-29 11:02:43.328289] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:37.984 [2024-10-29 11:02:43.330543] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:37.984 [2024-10-29 11:02:43.330588] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:37.984 [2024-10-29 11:02:43.330678] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:37.984 [2024-10-29 11:02:43.330718] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:37.984 [2024-10-29 11:02:43.330838] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:37.984 [2024-10-29 11:02:43.330972] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:37.984 spare 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.984 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:37.984 [2024-10-29 11:02:43.430896] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:13:37.984 [2024-10-29 11:02:43.430934] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:37.984 [2024-10-29 11:02:43.431268] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000036fc0 00:13:37.985 [2024-10-29 11:02:43.431512] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:13:37.985 [2024-10-29 11:02:43.431534] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:13:37.985 [2024-10-29 11:02:43.431714] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.985 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.245 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.245 "name": "raid_bdev1", 00:13:38.245 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:38.245 "strip_size_kb": 0, 00:13:38.245 "state": "online", 00:13:38.245 "raid_level": "raid1", 00:13:38.245 "superblock": true, 00:13:38.245 "num_base_bdevs": 4, 00:13:38.245 "num_base_bdevs_discovered": 3, 00:13:38.245 "num_base_bdevs_operational": 3, 00:13:38.245 "base_bdevs_list": [ 00:13:38.245 { 00:13:38.245 "name": "spare", 00:13:38.245 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:38.245 "is_configured": true, 00:13:38.245 "data_offset": 2048, 00:13:38.245 "data_size": 63488 00:13:38.245 }, 00:13:38.245 { 00:13:38.245 "name": null, 00:13:38.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.245 "is_configured": false, 00:13:38.245 "data_offset": 2048, 00:13:38.245 "data_size": 63488 00:13:38.245 }, 00:13:38.245 { 00:13:38.245 "name": "BaseBdev3", 00:13:38.245 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:38.245 "is_configured": true, 00:13:38.245 "data_offset": 2048, 00:13:38.245 "data_size": 63488 00:13:38.245 }, 00:13:38.245 { 00:13:38.245 "name": "BaseBdev4", 00:13:38.245 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:38.245 "is_configured": true, 00:13:38.245 "data_offset": 2048, 00:13:38.245 "data_size": 63488 00:13:38.245 } 00:13:38.245 ] 00:13:38.245 }' 00:13:38.245 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.245 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.505 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:38.505 "name": "raid_bdev1", 00:13:38.505 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:38.505 "strip_size_kb": 0, 00:13:38.505 "state": "online", 00:13:38.505 "raid_level": "raid1", 00:13:38.505 "superblock": true, 00:13:38.506 "num_base_bdevs": 4, 00:13:38.506 "num_base_bdevs_discovered": 3, 00:13:38.506 "num_base_bdevs_operational": 3, 00:13:38.506 "base_bdevs_list": [ 00:13:38.506 { 00:13:38.506 "name": "spare", 00:13:38.506 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:38.506 "is_configured": true, 00:13:38.506 "data_offset": 2048, 00:13:38.506 "data_size": 63488 00:13:38.506 }, 00:13:38.506 { 00:13:38.506 "name": null, 00:13:38.506 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.506 "is_configured": false, 00:13:38.506 "data_offset": 2048, 00:13:38.506 "data_size": 63488 00:13:38.506 }, 00:13:38.506 { 00:13:38.506 "name": "BaseBdev3", 00:13:38.506 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:38.506 "is_configured": true, 00:13:38.506 "data_offset": 2048, 00:13:38.506 "data_size": 63488 00:13:38.506 }, 00:13:38.506 { 00:13:38.506 "name": "BaseBdev4", 00:13:38.506 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:38.506 "is_configured": true, 00:13:38.506 "data_offset": 2048, 00:13:38.506 "data_size": 63488 00:13:38.506 } 00:13:38.506 ] 00:13:38.506 }' 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.506 11:02:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:38.506 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.766 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:38.766 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:38.766 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.766 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.767 [2024-10-29 11:02:44.035262] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.767 "name": "raid_bdev1", 00:13:38.767 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:38.767 "strip_size_kb": 0, 00:13:38.767 "state": "online", 00:13:38.767 "raid_level": "raid1", 00:13:38.767 "superblock": true, 00:13:38.767 "num_base_bdevs": 4, 00:13:38.767 "num_base_bdevs_discovered": 2, 00:13:38.767 "num_base_bdevs_operational": 2, 00:13:38.767 "base_bdevs_list": [ 00:13:38.767 { 00:13:38.767 "name": null, 00:13:38.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.767 "is_configured": false, 00:13:38.767 "data_offset": 0, 00:13:38.767 "data_size": 63488 00:13:38.767 }, 00:13:38.767 { 00:13:38.767 "name": null, 00:13:38.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.767 "is_configured": false, 00:13:38.767 "data_offset": 2048, 00:13:38.767 "data_size": 63488 00:13:38.767 }, 00:13:38.767 { 00:13:38.767 "name": "BaseBdev3", 00:13:38.767 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:38.767 "is_configured": true, 00:13:38.767 "data_offset": 2048, 00:13:38.767 "data_size": 63488 00:13:38.767 }, 00:13:38.767 { 00:13:38.767 "name": "BaseBdev4", 00:13:38.767 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:38.767 "is_configured": true, 00:13:38.767 "data_offset": 2048, 00:13:38.767 "data_size": 63488 00:13:38.767 } 00:13:38.767 ] 00:13:38.767 }' 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.767 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:39.027 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:39.027 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.027 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:39.027 [2024-10-29 11:02:44.486601] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:39.027 [2024-10-29 11:02:44.486818] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:39.027 [2024-10-29 11:02:44.486839] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:39.027 [2024-10-29 11:02:44.486888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:39.027 [2024-10-29 11:02:44.491565] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037090 00:13:39.027 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.027 11:02:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:39.027 [2024-10-29 11:02:44.493590] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:40.410 "name": "raid_bdev1", 00:13:40.410 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:40.410 "strip_size_kb": 0, 00:13:40.410 "state": "online", 00:13:40.410 "raid_level": "raid1", 00:13:40.410 "superblock": true, 00:13:40.410 "num_base_bdevs": 4, 00:13:40.410 "num_base_bdevs_discovered": 3, 00:13:40.410 "num_base_bdevs_operational": 3, 00:13:40.410 "process": { 00:13:40.410 "type": "rebuild", 00:13:40.410 "target": "spare", 00:13:40.410 "progress": { 00:13:40.410 "blocks": 20480, 00:13:40.410 "percent": 32 00:13:40.410 } 00:13:40.410 }, 00:13:40.410 "base_bdevs_list": [ 00:13:40.410 { 00:13:40.410 "name": "spare", 00:13:40.410 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:40.410 "is_configured": true, 00:13:40.410 "data_offset": 2048, 00:13:40.410 "data_size": 63488 00:13:40.410 }, 00:13:40.410 { 00:13:40.410 "name": null, 00:13:40.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.410 "is_configured": false, 00:13:40.410 "data_offset": 2048, 00:13:40.410 "data_size": 63488 00:13:40.410 }, 00:13:40.410 { 00:13:40.410 "name": "BaseBdev3", 00:13:40.410 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:40.410 "is_configured": true, 00:13:40.410 "data_offset": 2048, 00:13:40.410 "data_size": 63488 00:13:40.410 }, 00:13:40.410 { 00:13:40.410 "name": "BaseBdev4", 00:13:40.410 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:40.410 "is_configured": true, 00:13:40.410 "data_offset": 2048, 00:13:40.410 "data_size": 63488 00:13:40.410 } 00:13:40.410 ] 00:13:40.410 }' 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.410 [2024-10-29 11:02:45.657985] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:40.410 [2024-10-29 11:02:45.697829] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:40.410 [2024-10-29 11:02:45.697895] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:40.410 [2024-10-29 11:02:45.697929] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:40.410 [2024-10-29 11:02:45.697940] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.410 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:40.410 "name": "raid_bdev1", 00:13:40.410 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:40.410 "strip_size_kb": 0, 00:13:40.410 "state": "online", 00:13:40.410 "raid_level": "raid1", 00:13:40.410 "superblock": true, 00:13:40.410 "num_base_bdevs": 4, 00:13:40.410 "num_base_bdevs_discovered": 2, 00:13:40.410 "num_base_bdevs_operational": 2, 00:13:40.410 "base_bdevs_list": [ 00:13:40.410 { 00:13:40.410 "name": null, 00:13:40.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.410 "is_configured": false, 00:13:40.410 "data_offset": 0, 00:13:40.410 "data_size": 63488 00:13:40.410 }, 00:13:40.410 { 00:13:40.410 "name": null, 00:13:40.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:40.410 "is_configured": false, 00:13:40.410 "data_offset": 2048, 00:13:40.410 "data_size": 63488 00:13:40.410 }, 00:13:40.410 { 00:13:40.410 "name": "BaseBdev3", 00:13:40.411 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:40.411 "is_configured": true, 00:13:40.411 "data_offset": 2048, 00:13:40.411 "data_size": 63488 00:13:40.411 }, 00:13:40.411 { 00:13:40.411 "name": "BaseBdev4", 00:13:40.411 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:40.411 "is_configured": true, 00:13:40.411 "data_offset": 2048, 00:13:40.411 "data_size": 63488 00:13:40.411 } 00:13:40.411 ] 00:13:40.411 }' 00:13:40.411 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:40.411 11:02:45 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.671 11:02:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:40.671 11:02:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.671 11:02:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.671 [2024-10-29 11:02:46.169705] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:40.671 [2024-10-29 11:02:46.169772] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:40.671 [2024-10-29 11:02:46.169800] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:13:40.671 [2024-10-29 11:02:46.169813] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:40.671 [2024-10-29 11:02:46.170301] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:40.671 [2024-10-29 11:02:46.170336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:40.671 [2024-10-29 11:02:46.170447] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:40.671 [2024-10-29 11:02:46.170485] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:40.671 [2024-10-29 11:02:46.170506] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:40.671 [2024-10-29 11:02:46.170544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:40.931 spare 00:13:40.931 [2024-10-29 11:02:46.174733] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037160 00:13:40.931 11:02:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.931 11:02:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:40.931 [2024-10-29 11:02:46.176701] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:41.879 "name": "raid_bdev1", 00:13:41.879 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:41.879 "strip_size_kb": 0, 00:13:41.879 "state": "online", 00:13:41.879 "raid_level": "raid1", 00:13:41.879 "superblock": true, 00:13:41.879 "num_base_bdevs": 4, 00:13:41.879 "num_base_bdevs_discovered": 3, 00:13:41.879 "num_base_bdevs_operational": 3, 00:13:41.879 "process": { 00:13:41.879 "type": "rebuild", 00:13:41.879 "target": "spare", 00:13:41.879 "progress": { 00:13:41.879 "blocks": 20480, 00:13:41.879 "percent": 32 00:13:41.879 } 00:13:41.879 }, 00:13:41.879 "base_bdevs_list": [ 00:13:41.879 { 00:13:41.879 "name": "spare", 00:13:41.879 "uuid": "e8f6bc36-f964-5dac-9c9c-27a8e24ab268", 00:13:41.879 "is_configured": true, 00:13:41.879 "data_offset": 2048, 00:13:41.879 "data_size": 63488 00:13:41.879 }, 00:13:41.879 { 00:13:41.879 "name": null, 00:13:41.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.879 "is_configured": false, 00:13:41.879 "data_offset": 2048, 00:13:41.879 "data_size": 63488 00:13:41.879 }, 00:13:41.879 { 00:13:41.879 "name": "BaseBdev3", 00:13:41.879 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:41.879 "is_configured": true, 00:13:41.879 "data_offset": 2048, 00:13:41.879 "data_size": 63488 00:13:41.879 }, 00:13:41.879 { 00:13:41.879 "name": "BaseBdev4", 00:13:41.879 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:41.879 "is_configured": true, 00:13:41.879 "data_offset": 2048, 00:13:41.879 "data_size": 63488 00:13:41.879 } 00:13:41.879 ] 00:13:41.879 }' 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.879 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.879 [2024-10-29 11:02:47.337083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:42.140 [2024-10-29 11:02:47.380944] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:42.140 [2024-10-29 11:02:47.381034] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:42.140 [2024-10-29 11:02:47.381054] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:42.140 [2024-10-29 11:02:47.381063] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.140 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.140 "name": "raid_bdev1", 00:13:42.140 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:42.140 "strip_size_kb": 0, 00:13:42.140 "state": "online", 00:13:42.140 "raid_level": "raid1", 00:13:42.140 "superblock": true, 00:13:42.140 "num_base_bdevs": 4, 00:13:42.140 "num_base_bdevs_discovered": 2, 00:13:42.140 "num_base_bdevs_operational": 2, 00:13:42.140 "base_bdevs_list": [ 00:13:42.140 { 00:13:42.140 "name": null, 00:13:42.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.140 "is_configured": false, 00:13:42.140 "data_offset": 0, 00:13:42.140 "data_size": 63488 00:13:42.140 }, 00:13:42.140 { 00:13:42.140 "name": null, 00:13:42.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.140 "is_configured": false, 00:13:42.140 "data_offset": 2048, 00:13:42.140 "data_size": 63488 00:13:42.140 }, 00:13:42.140 { 00:13:42.140 "name": "BaseBdev3", 00:13:42.140 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:42.140 "is_configured": true, 00:13:42.140 "data_offset": 2048, 00:13:42.140 "data_size": 63488 00:13:42.140 }, 00:13:42.140 { 00:13:42.140 "name": "BaseBdev4", 00:13:42.140 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:42.140 "is_configured": true, 00:13:42.140 "data_offset": 2048, 00:13:42.140 "data_size": 63488 00:13:42.141 } 00:13:42.141 ] 00:13:42.141 }' 00:13:42.141 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.141 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:42.401 "name": "raid_bdev1", 00:13:42.401 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:42.401 "strip_size_kb": 0, 00:13:42.401 "state": "online", 00:13:42.401 "raid_level": "raid1", 00:13:42.401 "superblock": true, 00:13:42.401 "num_base_bdevs": 4, 00:13:42.401 "num_base_bdevs_discovered": 2, 00:13:42.401 "num_base_bdevs_operational": 2, 00:13:42.401 "base_bdevs_list": [ 00:13:42.401 { 00:13:42.401 "name": null, 00:13:42.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.401 "is_configured": false, 00:13:42.401 "data_offset": 0, 00:13:42.401 "data_size": 63488 00:13:42.401 }, 00:13:42.401 { 00:13:42.401 "name": null, 00:13:42.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:42.401 "is_configured": false, 00:13:42.401 "data_offset": 2048, 00:13:42.401 "data_size": 63488 00:13:42.401 }, 00:13:42.401 { 00:13:42.401 "name": "BaseBdev3", 00:13:42.401 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:42.401 "is_configured": true, 00:13:42.401 "data_offset": 2048, 00:13:42.401 "data_size": 63488 00:13:42.401 }, 00:13:42.401 { 00:13:42.401 "name": "BaseBdev4", 00:13:42.401 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:42.401 "is_configured": true, 00:13:42.401 "data_offset": 2048, 00:13:42.401 "data_size": 63488 00:13:42.401 } 00:13:42.401 ] 00:13:42.401 }' 00:13:42.401 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.662 [2024-10-29 11:02:47.960704] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:42.662 [2024-10-29 11:02:47.960766] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:42.662 [2024-10-29 11:02:47.960791] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:13:42.662 [2024-10-29 11:02:47.960803] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:42.662 [2024-10-29 11:02:47.961268] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:42.662 [2024-10-29 11:02:47.961303] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:42.662 [2024-10-29 11:02:47.961407] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:42.662 [2024-10-29 11:02:47.961424] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:42.662 [2024-10-29 11:02:47.961440] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:42.662 [2024-10-29 11:02:47.961452] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:42.662 BaseBdev1 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.662 11:02:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.604 11:02:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.604 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.604 "name": "raid_bdev1", 00:13:43.604 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:43.604 "strip_size_kb": 0, 00:13:43.604 "state": "online", 00:13:43.604 "raid_level": "raid1", 00:13:43.604 "superblock": true, 00:13:43.604 "num_base_bdevs": 4, 00:13:43.604 "num_base_bdevs_discovered": 2, 00:13:43.604 "num_base_bdevs_operational": 2, 00:13:43.604 "base_bdevs_list": [ 00:13:43.604 { 00:13:43.604 "name": null, 00:13:43.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.604 "is_configured": false, 00:13:43.604 "data_offset": 0, 00:13:43.604 "data_size": 63488 00:13:43.604 }, 00:13:43.604 { 00:13:43.604 "name": null, 00:13:43.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.604 "is_configured": false, 00:13:43.604 "data_offset": 2048, 00:13:43.604 "data_size": 63488 00:13:43.604 }, 00:13:43.604 { 00:13:43.604 "name": "BaseBdev3", 00:13:43.604 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:43.604 "is_configured": true, 00:13:43.604 "data_offset": 2048, 00:13:43.604 "data_size": 63488 00:13:43.604 }, 00:13:43.604 { 00:13:43.604 "name": "BaseBdev4", 00:13:43.604 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:43.604 "is_configured": true, 00:13:43.604 "data_offset": 2048, 00:13:43.604 "data_size": 63488 00:13:43.604 } 00:13:43.604 ] 00:13:43.604 }' 00:13:43.604 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.604 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:44.175 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.175 "name": "raid_bdev1", 00:13:44.175 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:44.175 "strip_size_kb": 0, 00:13:44.175 "state": "online", 00:13:44.175 "raid_level": "raid1", 00:13:44.175 "superblock": true, 00:13:44.175 "num_base_bdevs": 4, 00:13:44.175 "num_base_bdevs_discovered": 2, 00:13:44.175 "num_base_bdevs_operational": 2, 00:13:44.175 "base_bdevs_list": [ 00:13:44.175 { 00:13:44.175 "name": null, 00:13:44.175 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.175 "is_configured": false, 00:13:44.175 "data_offset": 0, 00:13:44.175 "data_size": 63488 00:13:44.175 }, 00:13:44.175 { 00:13:44.175 "name": null, 00:13:44.175 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.175 "is_configured": false, 00:13:44.175 "data_offset": 2048, 00:13:44.175 "data_size": 63488 00:13:44.175 }, 00:13:44.175 { 00:13:44.175 "name": "BaseBdev3", 00:13:44.176 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:44.176 "is_configured": true, 00:13:44.176 "data_offset": 2048, 00:13:44.176 "data_size": 63488 00:13:44.176 }, 00:13:44.176 { 00:13:44.176 "name": "BaseBdev4", 00:13:44.176 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:44.176 "is_configured": true, 00:13:44.176 "data_offset": 2048, 00:13:44.176 "data_size": 63488 00:13:44.176 } 00:13:44.176 ] 00:13:44.176 }' 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:44.176 [2024-10-29 11:02:49.570314] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:44.176 [2024-10-29 11:02:49.570495] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:44.176 [2024-10-29 11:02:49.570522] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:44.176 request: 00:13:44.176 { 00:13:44.176 "base_bdev": "BaseBdev1", 00:13:44.176 "raid_bdev": "raid_bdev1", 00:13:44.176 "method": "bdev_raid_add_base_bdev", 00:13:44.176 "req_id": 1 00:13:44.176 } 00:13:44.176 Got JSON-RPC error response 00:13:44.176 response: 00:13:44.176 { 00:13:44.176 "code": -22, 00:13:44.176 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:44.176 } 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:44.176 11:02:49 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.116 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.376 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.376 "name": "raid_bdev1", 00:13:45.376 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:45.376 "strip_size_kb": 0, 00:13:45.376 "state": "online", 00:13:45.376 "raid_level": "raid1", 00:13:45.376 "superblock": true, 00:13:45.376 "num_base_bdevs": 4, 00:13:45.376 "num_base_bdevs_discovered": 2, 00:13:45.376 "num_base_bdevs_operational": 2, 00:13:45.376 "base_bdevs_list": [ 00:13:45.376 { 00:13:45.376 "name": null, 00:13:45.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.376 "is_configured": false, 00:13:45.376 "data_offset": 0, 00:13:45.376 "data_size": 63488 00:13:45.376 }, 00:13:45.376 { 00:13:45.376 "name": null, 00:13:45.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.376 "is_configured": false, 00:13:45.376 "data_offset": 2048, 00:13:45.376 "data_size": 63488 00:13:45.376 }, 00:13:45.376 { 00:13:45.376 "name": "BaseBdev3", 00:13:45.376 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:45.376 "is_configured": true, 00:13:45.376 "data_offset": 2048, 00:13:45.376 "data_size": 63488 00:13:45.376 }, 00:13:45.376 { 00:13:45.376 "name": "BaseBdev4", 00:13:45.376 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:45.376 "is_configured": true, 00:13:45.376 "data_offset": 2048, 00:13:45.376 "data_size": 63488 00:13:45.376 } 00:13:45.376 ] 00:13:45.376 }' 00:13:45.376 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.376 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.637 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:45.637 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:45.637 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:45.637 11:02:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.637 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:45.637 "name": "raid_bdev1", 00:13:45.637 "uuid": "8d2053fd-9df7-47ec-9d0e-b6e401967f55", 00:13:45.637 "strip_size_kb": 0, 00:13:45.637 "state": "online", 00:13:45.637 "raid_level": "raid1", 00:13:45.638 "superblock": true, 00:13:45.638 "num_base_bdevs": 4, 00:13:45.638 "num_base_bdevs_discovered": 2, 00:13:45.638 "num_base_bdevs_operational": 2, 00:13:45.638 "base_bdevs_list": [ 00:13:45.638 { 00:13:45.638 "name": null, 00:13:45.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.638 "is_configured": false, 00:13:45.638 "data_offset": 0, 00:13:45.638 "data_size": 63488 00:13:45.638 }, 00:13:45.638 { 00:13:45.638 "name": null, 00:13:45.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.638 "is_configured": false, 00:13:45.638 "data_offset": 2048, 00:13:45.638 "data_size": 63488 00:13:45.638 }, 00:13:45.638 { 00:13:45.638 "name": "BaseBdev3", 00:13:45.638 "uuid": "87238f57-9f14-5b6b-887f-6c4b1b027889", 00:13:45.638 "is_configured": true, 00:13:45.638 "data_offset": 2048, 00:13:45.638 "data_size": 63488 00:13:45.638 }, 00:13:45.638 { 00:13:45.638 "name": "BaseBdev4", 00:13:45.638 "uuid": "556c0687-a756-53c9-a3b8-782acf913971", 00:13:45.638 "is_configured": true, 00:13:45.638 "data_offset": 2048, 00:13:45.638 "data_size": 63488 00:13:45.638 } 00:13:45.638 ] 00:13:45.638 }' 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89861 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@952 -- # '[' -z 89861 ']' 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # kill -0 89861 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@957 -- # uname 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:13:45.638 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 89861 00:13:45.898 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:13:45.898 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:13:45.898 killing process with pid 89861 00:13:45.898 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@970 -- # echo 'killing process with pid 89861' 00:13:45.898 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@971 -- # kill 89861 00:13:45.898 Received shutdown signal, test time was about 17.534206 seconds 00:13:45.898 00:13:45.898 Latency(us) 00:13:45.898 [2024-10-29T11:02:51.396Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:45.898 [2024-10-29T11:02:51.396Z] =================================================================================================================== 00:13:45.898 [2024-10-29T11:02:51.396Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:45.898 [2024-10-29 11:02:51.150520] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:45.898 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@976 -- # wait 89861 00:13:45.898 [2024-10-29 11:02:51.150690] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:45.898 [2024-10-29 11:02:51.150766] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:45.898 [2024-10-29 11:02:51.150789] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:13:45.898 [2024-10-29 11:02:51.197240] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:46.159 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:46.159 00:13:46.159 real 0m19.414s 00:13:46.159 user 0m25.723s 00:13:46.159 sys 0m2.381s 00:13:46.159 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1128 -- # xtrace_disable 00:13:46.159 11:02:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:46.159 ************************************ 00:13:46.159 END TEST raid_rebuild_test_sb_io 00:13:46.159 ************************************ 00:13:46.159 11:02:51 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:46.160 11:02:51 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:46.160 11:02:51 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:13:46.160 11:02:51 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:13:46.160 11:02:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:46.160 ************************************ 00:13:46.160 START TEST raid5f_state_function_test 00:13:46.160 ************************************ 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid5f 3 false 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90566 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90566' 00:13:46.160 Process raid pid: 90566 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90566 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 90566 ']' 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:13:46.160 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:13:46.160 11:02:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.160 [2024-10-29 11:02:51.576267] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:13:46.160 [2024-10-29 11:02:51.576429] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:46.420 [2024-10-29 11:02:51.725270] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:46.420 [2024-10-29 11:02:51.750491] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:46.420 [2024-10-29 11:02:51.794604] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:46.420 [2024-10-29 11:02:51.794653] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.992 [2024-10-29 11:02:52.404892] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:46.992 [2024-10-29 11:02:52.404983] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:46.992 [2024-10-29 11:02:52.404995] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:46.992 [2024-10-29 11:02:52.405007] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:46.992 [2024-10-29 11:02:52.405017] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:46.992 [2024-10-29 11:02:52.405032] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.992 "name": "Existed_Raid", 00:13:46.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.992 "strip_size_kb": 64, 00:13:46.992 "state": "configuring", 00:13:46.992 "raid_level": "raid5f", 00:13:46.992 "superblock": false, 00:13:46.992 "num_base_bdevs": 3, 00:13:46.992 "num_base_bdevs_discovered": 0, 00:13:46.992 "num_base_bdevs_operational": 3, 00:13:46.992 "base_bdevs_list": [ 00:13:46.992 { 00:13:46.992 "name": "BaseBdev1", 00:13:46.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.992 "is_configured": false, 00:13:46.992 "data_offset": 0, 00:13:46.992 "data_size": 0 00:13:46.992 }, 00:13:46.992 { 00:13:46.992 "name": "BaseBdev2", 00:13:46.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.992 "is_configured": false, 00:13:46.992 "data_offset": 0, 00:13:46.992 "data_size": 0 00:13:46.992 }, 00:13:46.992 { 00:13:46.992 "name": "BaseBdev3", 00:13:46.992 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.992 "is_configured": false, 00:13:46.992 "data_offset": 0, 00:13:46.992 "data_size": 0 00:13:46.992 } 00:13:46.992 ] 00:13:46.992 }' 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.992 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 [2024-10-29 11:02:52.836049] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:47.562 [2024-10-29 11:02:52.836102] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 [2024-10-29 11:02:52.848046] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:47.562 [2024-10-29 11:02:52.848094] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:47.562 [2024-10-29 11:02:52.848104] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:47.562 [2024-10-29 11:02:52.848115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:47.562 [2024-10-29 11:02:52.848123] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:47.562 [2024-10-29 11:02:52.848134] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 [2024-10-29 11:02:52.869242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:47.562 BaseBdev1 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 [ 00:13:47.562 { 00:13:47.562 "name": "BaseBdev1", 00:13:47.562 "aliases": [ 00:13:47.562 "ba296d33-c6c8-4502-a2d3-ec4f63b3001f" 00:13:47.562 ], 00:13:47.562 "product_name": "Malloc disk", 00:13:47.562 "block_size": 512, 00:13:47.562 "num_blocks": 65536, 00:13:47.562 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:47.562 "assigned_rate_limits": { 00:13:47.562 "rw_ios_per_sec": 0, 00:13:47.562 "rw_mbytes_per_sec": 0, 00:13:47.562 "r_mbytes_per_sec": 0, 00:13:47.562 "w_mbytes_per_sec": 0 00:13:47.562 }, 00:13:47.562 "claimed": true, 00:13:47.562 "claim_type": "exclusive_write", 00:13:47.562 "zoned": false, 00:13:47.562 "supported_io_types": { 00:13:47.562 "read": true, 00:13:47.562 "write": true, 00:13:47.562 "unmap": true, 00:13:47.562 "flush": true, 00:13:47.562 "reset": true, 00:13:47.562 "nvme_admin": false, 00:13:47.562 "nvme_io": false, 00:13:47.562 "nvme_io_md": false, 00:13:47.562 "write_zeroes": true, 00:13:47.562 "zcopy": true, 00:13:47.562 "get_zone_info": false, 00:13:47.562 "zone_management": false, 00:13:47.562 "zone_append": false, 00:13:47.562 "compare": false, 00:13:47.562 "compare_and_write": false, 00:13:47.562 "abort": true, 00:13:47.562 "seek_hole": false, 00:13:47.562 "seek_data": false, 00:13:47.562 "copy": true, 00:13:47.562 "nvme_iov_md": false 00:13:47.562 }, 00:13:47.562 "memory_domains": [ 00:13:47.562 { 00:13:47.562 "dma_device_id": "system", 00:13:47.562 "dma_device_type": 1 00:13:47.562 }, 00:13:47.562 { 00:13:47.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:47.562 "dma_device_type": 2 00:13:47.562 } 00:13:47.562 ], 00:13:47.562 "driver_specific": {} 00:13:47.562 } 00:13:47.562 ] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.562 "name": "Existed_Raid", 00:13:47.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.562 "strip_size_kb": 64, 00:13:47.562 "state": "configuring", 00:13:47.562 "raid_level": "raid5f", 00:13:47.562 "superblock": false, 00:13:47.562 "num_base_bdevs": 3, 00:13:47.562 "num_base_bdevs_discovered": 1, 00:13:47.562 "num_base_bdevs_operational": 3, 00:13:47.562 "base_bdevs_list": [ 00:13:47.562 { 00:13:47.562 "name": "BaseBdev1", 00:13:47.562 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:47.562 "is_configured": true, 00:13:47.562 "data_offset": 0, 00:13:47.562 "data_size": 65536 00:13:47.562 }, 00:13:47.562 { 00:13:47.562 "name": "BaseBdev2", 00:13:47.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.562 "is_configured": false, 00:13:47.562 "data_offset": 0, 00:13:47.562 "data_size": 0 00:13:47.562 }, 00:13:47.562 { 00:13:47.562 "name": "BaseBdev3", 00:13:47.562 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.562 "is_configured": false, 00:13:47.562 "data_offset": 0, 00:13:47.562 "data_size": 0 00:13:47.562 } 00:13:47.562 ] 00:13:47.562 }' 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.562 11:02:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.822 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:47.822 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.822 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.081 [2024-10-29 11:02:53.324521] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:48.081 [2024-10-29 11:02:53.324572] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.081 [2024-10-29 11:02:53.332561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:48.081 [2024-10-29 11:02:53.334507] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:48.081 [2024-10-29 11:02:53.334550] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:48.081 [2024-10-29 11:02:53.334560] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:48.081 [2024-10-29 11:02:53.334573] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:48.081 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.082 "name": "Existed_Raid", 00:13:48.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.082 "strip_size_kb": 64, 00:13:48.082 "state": "configuring", 00:13:48.082 "raid_level": "raid5f", 00:13:48.082 "superblock": false, 00:13:48.082 "num_base_bdevs": 3, 00:13:48.082 "num_base_bdevs_discovered": 1, 00:13:48.082 "num_base_bdevs_operational": 3, 00:13:48.082 "base_bdevs_list": [ 00:13:48.082 { 00:13:48.082 "name": "BaseBdev1", 00:13:48.082 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:48.082 "is_configured": true, 00:13:48.082 "data_offset": 0, 00:13:48.082 "data_size": 65536 00:13:48.082 }, 00:13:48.082 { 00:13:48.082 "name": "BaseBdev2", 00:13:48.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.082 "is_configured": false, 00:13:48.082 "data_offset": 0, 00:13:48.082 "data_size": 0 00:13:48.082 }, 00:13:48.082 { 00:13:48.082 "name": "BaseBdev3", 00:13:48.082 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.082 "is_configured": false, 00:13:48.082 "data_offset": 0, 00:13:48.082 "data_size": 0 00:13:48.082 } 00:13:48.082 ] 00:13:48.082 }' 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.082 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.341 [2024-10-29 11:02:53.795169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:48.341 BaseBdev2 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.341 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.341 [ 00:13:48.341 { 00:13:48.341 "name": "BaseBdev2", 00:13:48.342 "aliases": [ 00:13:48.342 "2e79dff1-34b6-4b90-a305-978712dfcf60" 00:13:48.342 ], 00:13:48.342 "product_name": "Malloc disk", 00:13:48.342 "block_size": 512, 00:13:48.342 "num_blocks": 65536, 00:13:48.342 "uuid": "2e79dff1-34b6-4b90-a305-978712dfcf60", 00:13:48.342 "assigned_rate_limits": { 00:13:48.342 "rw_ios_per_sec": 0, 00:13:48.342 "rw_mbytes_per_sec": 0, 00:13:48.342 "r_mbytes_per_sec": 0, 00:13:48.342 "w_mbytes_per_sec": 0 00:13:48.342 }, 00:13:48.342 "claimed": true, 00:13:48.342 "claim_type": "exclusive_write", 00:13:48.342 "zoned": false, 00:13:48.342 "supported_io_types": { 00:13:48.342 "read": true, 00:13:48.342 "write": true, 00:13:48.342 "unmap": true, 00:13:48.342 "flush": true, 00:13:48.342 "reset": true, 00:13:48.342 "nvme_admin": false, 00:13:48.342 "nvme_io": false, 00:13:48.342 "nvme_io_md": false, 00:13:48.342 "write_zeroes": true, 00:13:48.342 "zcopy": true, 00:13:48.342 "get_zone_info": false, 00:13:48.342 "zone_management": false, 00:13:48.342 "zone_append": false, 00:13:48.342 "compare": false, 00:13:48.342 "compare_and_write": false, 00:13:48.342 "abort": true, 00:13:48.342 "seek_hole": false, 00:13:48.342 "seek_data": false, 00:13:48.342 "copy": true, 00:13:48.342 "nvme_iov_md": false 00:13:48.342 }, 00:13:48.342 "memory_domains": [ 00:13:48.342 { 00:13:48.342 "dma_device_id": "system", 00:13:48.342 "dma_device_type": 1 00:13:48.342 }, 00:13:48.342 { 00:13:48.342 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:48.342 "dma_device_type": 2 00:13:48.342 } 00:13:48.342 ], 00:13:48.342 "driver_specific": {} 00:13:48.342 } 00:13:48.342 ] 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.342 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.602 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.602 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.602 "name": "Existed_Raid", 00:13:48.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.602 "strip_size_kb": 64, 00:13:48.602 "state": "configuring", 00:13:48.602 "raid_level": "raid5f", 00:13:48.602 "superblock": false, 00:13:48.602 "num_base_bdevs": 3, 00:13:48.602 "num_base_bdevs_discovered": 2, 00:13:48.602 "num_base_bdevs_operational": 3, 00:13:48.602 "base_bdevs_list": [ 00:13:48.602 { 00:13:48.602 "name": "BaseBdev1", 00:13:48.602 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:48.602 "is_configured": true, 00:13:48.602 "data_offset": 0, 00:13:48.602 "data_size": 65536 00:13:48.602 }, 00:13:48.602 { 00:13:48.602 "name": "BaseBdev2", 00:13:48.602 "uuid": "2e79dff1-34b6-4b90-a305-978712dfcf60", 00:13:48.602 "is_configured": true, 00:13:48.602 "data_offset": 0, 00:13:48.602 "data_size": 65536 00:13:48.602 }, 00:13:48.602 { 00:13:48.602 "name": "BaseBdev3", 00:13:48.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.602 "is_configured": false, 00:13:48.602 "data_offset": 0, 00:13:48.602 "data_size": 0 00:13:48.602 } 00:13:48.602 ] 00:13:48.602 }' 00:13:48.602 11:02:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.602 11:02:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.863 [2024-10-29 11:02:54.300407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:48.863 [2024-10-29 11:02:54.300618] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:13:48.863 [2024-10-29 11:02:54.300676] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:48.863 [2024-10-29 11:02:54.301732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:13:48.863 [2024-10-29 11:02:54.303424] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:13:48.863 [2024-10-29 11:02:54.303492] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:13:48.863 BaseBdev3 00:13:48.863 [2024-10-29 11:02:54.304291] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.863 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.863 [ 00:13:48.863 { 00:13:48.863 "name": "BaseBdev3", 00:13:48.863 "aliases": [ 00:13:48.863 "d86cc0d6-dbbe-42ef-891f-375229a8d01d" 00:13:48.863 ], 00:13:48.863 "product_name": "Malloc disk", 00:13:48.863 "block_size": 512, 00:13:48.863 "num_blocks": 65536, 00:13:48.863 "uuid": "d86cc0d6-dbbe-42ef-891f-375229a8d01d", 00:13:48.863 "assigned_rate_limits": { 00:13:48.863 "rw_ios_per_sec": 0, 00:13:48.863 "rw_mbytes_per_sec": 0, 00:13:48.863 "r_mbytes_per_sec": 0, 00:13:48.863 "w_mbytes_per_sec": 0 00:13:48.863 }, 00:13:48.863 "claimed": true, 00:13:48.863 "claim_type": "exclusive_write", 00:13:48.863 "zoned": false, 00:13:48.863 "supported_io_types": { 00:13:48.863 "read": true, 00:13:48.863 "write": true, 00:13:48.863 "unmap": true, 00:13:48.863 "flush": true, 00:13:48.863 "reset": true, 00:13:48.863 "nvme_admin": false, 00:13:48.863 "nvme_io": false, 00:13:48.863 "nvme_io_md": false, 00:13:48.863 "write_zeroes": true, 00:13:48.863 "zcopy": true, 00:13:48.863 "get_zone_info": false, 00:13:48.863 "zone_management": false, 00:13:48.863 "zone_append": false, 00:13:48.863 "compare": false, 00:13:48.863 "compare_and_write": false, 00:13:48.863 "abort": true, 00:13:48.863 "seek_hole": false, 00:13:48.863 "seek_data": false, 00:13:48.863 "copy": true, 00:13:48.863 "nvme_iov_md": false 00:13:48.863 }, 00:13:48.863 "memory_domains": [ 00:13:48.863 { 00:13:48.864 "dma_device_id": "system", 00:13:48.864 "dma_device_type": 1 00:13:48.864 }, 00:13:48.864 { 00:13:48.864 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:48.864 "dma_device_type": 2 00:13:48.864 } 00:13:48.864 ], 00:13:48.864 "driver_specific": {} 00:13:48.864 } 00:13:48.864 ] 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.864 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.124 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.124 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.124 "name": "Existed_Raid", 00:13:49.124 "uuid": "9f57dc90-b071-412e-be0e-eff0ad24db4a", 00:13:49.124 "strip_size_kb": 64, 00:13:49.124 "state": "online", 00:13:49.124 "raid_level": "raid5f", 00:13:49.124 "superblock": false, 00:13:49.124 "num_base_bdevs": 3, 00:13:49.124 "num_base_bdevs_discovered": 3, 00:13:49.124 "num_base_bdevs_operational": 3, 00:13:49.124 "base_bdevs_list": [ 00:13:49.124 { 00:13:49.124 "name": "BaseBdev1", 00:13:49.124 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:49.124 "is_configured": true, 00:13:49.124 "data_offset": 0, 00:13:49.124 "data_size": 65536 00:13:49.124 }, 00:13:49.124 { 00:13:49.124 "name": "BaseBdev2", 00:13:49.124 "uuid": "2e79dff1-34b6-4b90-a305-978712dfcf60", 00:13:49.124 "is_configured": true, 00:13:49.124 "data_offset": 0, 00:13:49.124 "data_size": 65536 00:13:49.124 }, 00:13:49.124 { 00:13:49.124 "name": "BaseBdev3", 00:13:49.124 "uuid": "d86cc0d6-dbbe-42ef-891f-375229a8d01d", 00:13:49.124 "is_configured": true, 00:13:49.124 "data_offset": 0, 00:13:49.124 "data_size": 65536 00:13:49.124 } 00:13:49.124 ] 00:13:49.124 }' 00:13:49.124 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.124 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:49.385 [2024-10-29 11:02:54.823986] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:49.385 "name": "Existed_Raid", 00:13:49.385 "aliases": [ 00:13:49.385 "9f57dc90-b071-412e-be0e-eff0ad24db4a" 00:13:49.385 ], 00:13:49.385 "product_name": "Raid Volume", 00:13:49.385 "block_size": 512, 00:13:49.385 "num_blocks": 131072, 00:13:49.385 "uuid": "9f57dc90-b071-412e-be0e-eff0ad24db4a", 00:13:49.385 "assigned_rate_limits": { 00:13:49.385 "rw_ios_per_sec": 0, 00:13:49.385 "rw_mbytes_per_sec": 0, 00:13:49.385 "r_mbytes_per_sec": 0, 00:13:49.385 "w_mbytes_per_sec": 0 00:13:49.385 }, 00:13:49.385 "claimed": false, 00:13:49.385 "zoned": false, 00:13:49.385 "supported_io_types": { 00:13:49.385 "read": true, 00:13:49.385 "write": true, 00:13:49.385 "unmap": false, 00:13:49.385 "flush": false, 00:13:49.385 "reset": true, 00:13:49.385 "nvme_admin": false, 00:13:49.385 "nvme_io": false, 00:13:49.385 "nvme_io_md": false, 00:13:49.385 "write_zeroes": true, 00:13:49.385 "zcopy": false, 00:13:49.385 "get_zone_info": false, 00:13:49.385 "zone_management": false, 00:13:49.385 "zone_append": false, 00:13:49.385 "compare": false, 00:13:49.385 "compare_and_write": false, 00:13:49.385 "abort": false, 00:13:49.385 "seek_hole": false, 00:13:49.385 "seek_data": false, 00:13:49.385 "copy": false, 00:13:49.385 "nvme_iov_md": false 00:13:49.385 }, 00:13:49.385 "driver_specific": { 00:13:49.385 "raid": { 00:13:49.385 "uuid": "9f57dc90-b071-412e-be0e-eff0ad24db4a", 00:13:49.385 "strip_size_kb": 64, 00:13:49.385 "state": "online", 00:13:49.385 "raid_level": "raid5f", 00:13:49.385 "superblock": false, 00:13:49.385 "num_base_bdevs": 3, 00:13:49.385 "num_base_bdevs_discovered": 3, 00:13:49.385 "num_base_bdevs_operational": 3, 00:13:49.385 "base_bdevs_list": [ 00:13:49.385 { 00:13:49.385 "name": "BaseBdev1", 00:13:49.385 "uuid": "ba296d33-c6c8-4502-a2d3-ec4f63b3001f", 00:13:49.385 "is_configured": true, 00:13:49.385 "data_offset": 0, 00:13:49.385 "data_size": 65536 00:13:49.385 }, 00:13:49.385 { 00:13:49.385 "name": "BaseBdev2", 00:13:49.385 "uuid": "2e79dff1-34b6-4b90-a305-978712dfcf60", 00:13:49.385 "is_configured": true, 00:13:49.385 "data_offset": 0, 00:13:49.385 "data_size": 65536 00:13:49.385 }, 00:13:49.385 { 00:13:49.385 "name": "BaseBdev3", 00:13:49.385 "uuid": "d86cc0d6-dbbe-42ef-891f-375229a8d01d", 00:13:49.385 "is_configured": true, 00:13:49.385 "data_offset": 0, 00:13:49.385 "data_size": 65536 00:13:49.385 } 00:13:49.385 ] 00:13:49.385 } 00:13:49.385 } 00:13:49.385 }' 00:13:49.385 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:49.645 BaseBdev2 00:13:49.645 BaseBdev3' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.645 11:02:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.645 [2024-10-29 11:02:55.083342] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.645 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.905 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.905 "name": "Existed_Raid", 00:13:49.905 "uuid": "9f57dc90-b071-412e-be0e-eff0ad24db4a", 00:13:49.905 "strip_size_kb": 64, 00:13:49.905 "state": "online", 00:13:49.905 "raid_level": "raid5f", 00:13:49.905 "superblock": false, 00:13:49.905 "num_base_bdevs": 3, 00:13:49.905 "num_base_bdevs_discovered": 2, 00:13:49.905 "num_base_bdevs_operational": 2, 00:13:49.905 "base_bdevs_list": [ 00:13:49.905 { 00:13:49.905 "name": null, 00:13:49.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:49.905 "is_configured": false, 00:13:49.905 "data_offset": 0, 00:13:49.905 "data_size": 65536 00:13:49.905 }, 00:13:49.905 { 00:13:49.905 "name": "BaseBdev2", 00:13:49.905 "uuid": "2e79dff1-34b6-4b90-a305-978712dfcf60", 00:13:49.905 "is_configured": true, 00:13:49.905 "data_offset": 0, 00:13:49.905 "data_size": 65536 00:13:49.905 }, 00:13:49.905 { 00:13:49.905 "name": "BaseBdev3", 00:13:49.905 "uuid": "d86cc0d6-dbbe-42ef-891f-375229a8d01d", 00:13:49.905 "is_configured": true, 00:13:49.905 "data_offset": 0, 00:13:49.905 "data_size": 65536 00:13:49.905 } 00:13:49.905 ] 00:13:49.905 }' 00:13:49.905 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.905 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.164 [2024-10-29 11:02:55.569907] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:50.164 [2024-10-29 11:02:55.570021] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:50.164 [2024-10-29 11:02:55.581309] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.164 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.165 [2024-10-29 11:02:55.641238] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:50.165 [2024-10-29 11:02:55.641291] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.165 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.424 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.424 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 BaseBdev2 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 [ 00:13:50.425 { 00:13:50.425 "name": "BaseBdev2", 00:13:50.425 "aliases": [ 00:13:50.425 "626e691a-e973-41a6-9904-111e9d200293" 00:13:50.425 ], 00:13:50.425 "product_name": "Malloc disk", 00:13:50.425 "block_size": 512, 00:13:50.425 "num_blocks": 65536, 00:13:50.425 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:50.425 "assigned_rate_limits": { 00:13:50.425 "rw_ios_per_sec": 0, 00:13:50.425 "rw_mbytes_per_sec": 0, 00:13:50.425 "r_mbytes_per_sec": 0, 00:13:50.425 "w_mbytes_per_sec": 0 00:13:50.425 }, 00:13:50.425 "claimed": false, 00:13:50.425 "zoned": false, 00:13:50.425 "supported_io_types": { 00:13:50.425 "read": true, 00:13:50.425 "write": true, 00:13:50.425 "unmap": true, 00:13:50.425 "flush": true, 00:13:50.425 "reset": true, 00:13:50.425 "nvme_admin": false, 00:13:50.425 "nvme_io": false, 00:13:50.425 "nvme_io_md": false, 00:13:50.425 "write_zeroes": true, 00:13:50.425 "zcopy": true, 00:13:50.425 "get_zone_info": false, 00:13:50.425 "zone_management": false, 00:13:50.425 "zone_append": false, 00:13:50.425 "compare": false, 00:13:50.425 "compare_and_write": false, 00:13:50.425 "abort": true, 00:13:50.425 "seek_hole": false, 00:13:50.425 "seek_data": false, 00:13:50.425 "copy": true, 00:13:50.425 "nvme_iov_md": false 00:13:50.425 }, 00:13:50.425 "memory_domains": [ 00:13:50.425 { 00:13:50.425 "dma_device_id": "system", 00:13:50.425 "dma_device_type": 1 00:13:50.425 }, 00:13:50.425 { 00:13:50.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.425 "dma_device_type": 2 00:13:50.425 } 00:13:50.425 ], 00:13:50.425 "driver_specific": {} 00:13:50.425 } 00:13:50.425 ] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 BaseBdev3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 [ 00:13:50.425 { 00:13:50.425 "name": "BaseBdev3", 00:13:50.425 "aliases": [ 00:13:50.425 "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e" 00:13:50.425 ], 00:13:50.425 "product_name": "Malloc disk", 00:13:50.425 "block_size": 512, 00:13:50.425 "num_blocks": 65536, 00:13:50.425 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:50.425 "assigned_rate_limits": { 00:13:50.425 "rw_ios_per_sec": 0, 00:13:50.425 "rw_mbytes_per_sec": 0, 00:13:50.425 "r_mbytes_per_sec": 0, 00:13:50.425 "w_mbytes_per_sec": 0 00:13:50.425 }, 00:13:50.425 "claimed": false, 00:13:50.425 "zoned": false, 00:13:50.425 "supported_io_types": { 00:13:50.425 "read": true, 00:13:50.425 "write": true, 00:13:50.425 "unmap": true, 00:13:50.425 "flush": true, 00:13:50.425 "reset": true, 00:13:50.425 "nvme_admin": false, 00:13:50.425 "nvme_io": false, 00:13:50.425 "nvme_io_md": false, 00:13:50.425 "write_zeroes": true, 00:13:50.425 "zcopy": true, 00:13:50.425 "get_zone_info": false, 00:13:50.425 "zone_management": false, 00:13:50.425 "zone_append": false, 00:13:50.425 "compare": false, 00:13:50.425 "compare_and_write": false, 00:13:50.425 "abort": true, 00:13:50.425 "seek_hole": false, 00:13:50.425 "seek_data": false, 00:13:50.425 "copy": true, 00:13:50.425 "nvme_iov_md": false 00:13:50.425 }, 00:13:50.425 "memory_domains": [ 00:13:50.425 { 00:13:50.425 "dma_device_id": "system", 00:13:50.425 "dma_device_type": 1 00:13:50.425 }, 00:13:50.425 { 00:13:50.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.425 "dma_device_type": 2 00:13:50.425 } 00:13:50.425 ], 00:13:50.425 "driver_specific": {} 00:13:50.425 } 00:13:50.425 ] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 [2024-10-29 11:02:55.812251] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:50.425 [2024-10-29 11:02:55.812349] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:50.425 [2024-10-29 11:02:55.812427] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:50.425 [2024-10-29 11:02:55.814233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.425 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.425 "name": "Existed_Raid", 00:13:50.425 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.425 "strip_size_kb": 64, 00:13:50.425 "state": "configuring", 00:13:50.425 "raid_level": "raid5f", 00:13:50.425 "superblock": false, 00:13:50.425 "num_base_bdevs": 3, 00:13:50.425 "num_base_bdevs_discovered": 2, 00:13:50.426 "num_base_bdevs_operational": 3, 00:13:50.426 "base_bdevs_list": [ 00:13:50.426 { 00:13:50.426 "name": "BaseBdev1", 00:13:50.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.426 "is_configured": false, 00:13:50.426 "data_offset": 0, 00:13:50.426 "data_size": 0 00:13:50.426 }, 00:13:50.426 { 00:13:50.426 "name": "BaseBdev2", 00:13:50.426 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:50.426 "is_configured": true, 00:13:50.426 "data_offset": 0, 00:13:50.426 "data_size": 65536 00:13:50.426 }, 00:13:50.426 { 00:13:50.426 "name": "BaseBdev3", 00:13:50.426 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:50.426 "is_configured": true, 00:13:50.426 "data_offset": 0, 00:13:50.426 "data_size": 65536 00:13:50.426 } 00:13:50.426 ] 00:13:50.426 }' 00:13:50.426 11:02:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.426 11:02:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.995 [2024-10-29 11:02:56.235559] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.995 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.996 "name": "Existed_Raid", 00:13:50.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.996 "strip_size_kb": 64, 00:13:50.996 "state": "configuring", 00:13:50.996 "raid_level": "raid5f", 00:13:50.996 "superblock": false, 00:13:50.996 "num_base_bdevs": 3, 00:13:50.996 "num_base_bdevs_discovered": 1, 00:13:50.996 "num_base_bdevs_operational": 3, 00:13:50.996 "base_bdevs_list": [ 00:13:50.996 { 00:13:50.996 "name": "BaseBdev1", 00:13:50.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.996 "is_configured": false, 00:13:50.996 "data_offset": 0, 00:13:50.996 "data_size": 0 00:13:50.996 }, 00:13:50.996 { 00:13:50.996 "name": null, 00:13:50.996 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:50.996 "is_configured": false, 00:13:50.996 "data_offset": 0, 00:13:50.996 "data_size": 65536 00:13:50.996 }, 00:13:50.996 { 00:13:50.996 "name": "BaseBdev3", 00:13:50.996 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:50.996 "is_configured": true, 00:13:50.996 "data_offset": 0, 00:13:50.996 "data_size": 65536 00:13:50.996 } 00:13:50.996 ] 00:13:50.996 }' 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.996 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.256 [2024-10-29 11:02:56.718113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:51.256 BaseBdev1 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.256 [ 00:13:51.256 { 00:13:51.256 "name": "BaseBdev1", 00:13:51.256 "aliases": [ 00:13:51.256 "3815fc25-b0b9-4fe1-8644-b2d017f9aa63" 00:13:51.256 ], 00:13:51.256 "product_name": "Malloc disk", 00:13:51.256 "block_size": 512, 00:13:51.256 "num_blocks": 65536, 00:13:51.256 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:51.256 "assigned_rate_limits": { 00:13:51.256 "rw_ios_per_sec": 0, 00:13:51.256 "rw_mbytes_per_sec": 0, 00:13:51.256 "r_mbytes_per_sec": 0, 00:13:51.256 "w_mbytes_per_sec": 0 00:13:51.256 }, 00:13:51.256 "claimed": true, 00:13:51.256 "claim_type": "exclusive_write", 00:13:51.256 "zoned": false, 00:13:51.256 "supported_io_types": { 00:13:51.256 "read": true, 00:13:51.256 "write": true, 00:13:51.256 "unmap": true, 00:13:51.256 "flush": true, 00:13:51.256 "reset": true, 00:13:51.256 "nvme_admin": false, 00:13:51.256 "nvme_io": false, 00:13:51.256 "nvme_io_md": false, 00:13:51.256 "write_zeroes": true, 00:13:51.256 "zcopy": true, 00:13:51.256 "get_zone_info": false, 00:13:51.256 "zone_management": false, 00:13:51.256 "zone_append": false, 00:13:51.256 "compare": false, 00:13:51.256 "compare_and_write": false, 00:13:51.256 "abort": true, 00:13:51.256 "seek_hole": false, 00:13:51.256 "seek_data": false, 00:13:51.256 "copy": true, 00:13:51.256 "nvme_iov_md": false 00:13:51.256 }, 00:13:51.256 "memory_domains": [ 00:13:51.256 { 00:13:51.256 "dma_device_id": "system", 00:13:51.256 "dma_device_type": 1 00:13:51.256 }, 00:13:51.256 { 00:13:51.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:51.256 "dma_device_type": 2 00:13:51.256 } 00:13:51.256 ], 00:13:51.256 "driver_specific": {} 00:13:51.256 } 00:13:51.256 ] 00:13:51.256 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.517 "name": "Existed_Raid", 00:13:51.517 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.517 "strip_size_kb": 64, 00:13:51.517 "state": "configuring", 00:13:51.517 "raid_level": "raid5f", 00:13:51.517 "superblock": false, 00:13:51.517 "num_base_bdevs": 3, 00:13:51.517 "num_base_bdevs_discovered": 2, 00:13:51.517 "num_base_bdevs_operational": 3, 00:13:51.517 "base_bdevs_list": [ 00:13:51.517 { 00:13:51.517 "name": "BaseBdev1", 00:13:51.517 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:51.517 "is_configured": true, 00:13:51.517 "data_offset": 0, 00:13:51.517 "data_size": 65536 00:13:51.517 }, 00:13:51.517 { 00:13:51.517 "name": null, 00:13:51.517 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:51.517 "is_configured": false, 00:13:51.517 "data_offset": 0, 00:13:51.517 "data_size": 65536 00:13:51.517 }, 00:13:51.517 { 00:13:51.517 "name": "BaseBdev3", 00:13:51.517 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:51.517 "is_configured": true, 00:13:51.517 "data_offset": 0, 00:13:51.517 "data_size": 65536 00:13:51.517 } 00:13:51.517 ] 00:13:51.517 }' 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.517 11:02:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.778 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.039 [2024-10-29 11:02:57.277303] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.039 "name": "Existed_Raid", 00:13:52.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.039 "strip_size_kb": 64, 00:13:52.039 "state": "configuring", 00:13:52.039 "raid_level": "raid5f", 00:13:52.039 "superblock": false, 00:13:52.039 "num_base_bdevs": 3, 00:13:52.039 "num_base_bdevs_discovered": 1, 00:13:52.039 "num_base_bdevs_operational": 3, 00:13:52.039 "base_bdevs_list": [ 00:13:52.039 { 00:13:52.039 "name": "BaseBdev1", 00:13:52.039 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:52.039 "is_configured": true, 00:13:52.039 "data_offset": 0, 00:13:52.039 "data_size": 65536 00:13:52.039 }, 00:13:52.039 { 00:13:52.039 "name": null, 00:13:52.039 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:52.039 "is_configured": false, 00:13:52.039 "data_offset": 0, 00:13:52.039 "data_size": 65536 00:13:52.039 }, 00:13:52.039 { 00:13:52.039 "name": null, 00:13:52.039 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:52.039 "is_configured": false, 00:13:52.039 "data_offset": 0, 00:13:52.039 "data_size": 65536 00:13:52.039 } 00:13:52.039 ] 00:13:52.039 }' 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.039 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.299 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.299 [2024-10-29 11:02:57.796456] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.559 "name": "Existed_Raid", 00:13:52.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.559 "strip_size_kb": 64, 00:13:52.559 "state": "configuring", 00:13:52.559 "raid_level": "raid5f", 00:13:52.559 "superblock": false, 00:13:52.559 "num_base_bdevs": 3, 00:13:52.559 "num_base_bdevs_discovered": 2, 00:13:52.559 "num_base_bdevs_operational": 3, 00:13:52.559 "base_bdevs_list": [ 00:13:52.559 { 00:13:52.559 "name": "BaseBdev1", 00:13:52.559 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:52.559 "is_configured": true, 00:13:52.559 "data_offset": 0, 00:13:52.559 "data_size": 65536 00:13:52.559 }, 00:13:52.559 { 00:13:52.559 "name": null, 00:13:52.559 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:52.559 "is_configured": false, 00:13:52.559 "data_offset": 0, 00:13:52.559 "data_size": 65536 00:13:52.559 }, 00:13:52.559 { 00:13:52.559 "name": "BaseBdev3", 00:13:52.559 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:52.559 "is_configured": true, 00:13:52.559 "data_offset": 0, 00:13:52.559 "data_size": 65536 00:13:52.559 } 00:13:52.559 ] 00:13:52.559 }' 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.559 11:02:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.819 [2024-10-29 11:02:58.299718] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.819 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.078 "name": "Existed_Raid", 00:13:53.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.078 "strip_size_kb": 64, 00:13:53.078 "state": "configuring", 00:13:53.078 "raid_level": "raid5f", 00:13:53.078 "superblock": false, 00:13:53.078 "num_base_bdevs": 3, 00:13:53.078 "num_base_bdevs_discovered": 1, 00:13:53.078 "num_base_bdevs_operational": 3, 00:13:53.078 "base_bdevs_list": [ 00:13:53.078 { 00:13:53.078 "name": null, 00:13:53.078 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:53.078 "is_configured": false, 00:13:53.078 "data_offset": 0, 00:13:53.078 "data_size": 65536 00:13:53.078 }, 00:13:53.078 { 00:13:53.078 "name": null, 00:13:53.078 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:53.078 "is_configured": false, 00:13:53.078 "data_offset": 0, 00:13:53.078 "data_size": 65536 00:13:53.078 }, 00:13:53.078 { 00:13:53.078 "name": "BaseBdev3", 00:13:53.078 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:53.078 "is_configured": true, 00:13:53.078 "data_offset": 0, 00:13:53.078 "data_size": 65536 00:13:53.078 } 00:13:53.078 ] 00:13:53.078 }' 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.078 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.337 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.597 [2024-10-29 11:02:58.841407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.597 "name": "Existed_Raid", 00:13:53.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.597 "strip_size_kb": 64, 00:13:53.597 "state": "configuring", 00:13:53.597 "raid_level": "raid5f", 00:13:53.597 "superblock": false, 00:13:53.597 "num_base_bdevs": 3, 00:13:53.597 "num_base_bdevs_discovered": 2, 00:13:53.597 "num_base_bdevs_operational": 3, 00:13:53.597 "base_bdevs_list": [ 00:13:53.597 { 00:13:53.597 "name": null, 00:13:53.597 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:53.597 "is_configured": false, 00:13:53.597 "data_offset": 0, 00:13:53.597 "data_size": 65536 00:13:53.597 }, 00:13:53.597 { 00:13:53.597 "name": "BaseBdev2", 00:13:53.597 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:53.597 "is_configured": true, 00:13:53.597 "data_offset": 0, 00:13:53.597 "data_size": 65536 00:13:53.597 }, 00:13:53.597 { 00:13:53.597 "name": "BaseBdev3", 00:13:53.597 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:53.597 "is_configured": true, 00:13:53.597 "data_offset": 0, 00:13:53.597 "data_size": 65536 00:13:53.597 } 00:13:53.597 ] 00:13:53.597 }' 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.597 11:02:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3815fc25-b0b9-4fe1-8644-b2d017f9aa63 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.857 [2024-10-29 11:02:59.347723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:53.857 [2024-10-29 11:02:59.347853] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:13:53.857 [2024-10-29 11:02:59.347887] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:53.857 [2024-10-29 11:02:59.348220] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:13:53.857 [2024-10-29 11:02:59.348733] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:13:53.857 [2024-10-29 11:02:59.348790] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:13:53.857 [2024-10-29 11:02:59.349062] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:53.857 NewBaseBdev 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.857 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.117 [ 00:13:54.117 { 00:13:54.117 "name": "NewBaseBdev", 00:13:54.117 "aliases": [ 00:13:54.117 "3815fc25-b0b9-4fe1-8644-b2d017f9aa63" 00:13:54.117 ], 00:13:54.117 "product_name": "Malloc disk", 00:13:54.117 "block_size": 512, 00:13:54.117 "num_blocks": 65536, 00:13:54.117 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:54.117 "assigned_rate_limits": { 00:13:54.117 "rw_ios_per_sec": 0, 00:13:54.117 "rw_mbytes_per_sec": 0, 00:13:54.117 "r_mbytes_per_sec": 0, 00:13:54.117 "w_mbytes_per_sec": 0 00:13:54.117 }, 00:13:54.117 "claimed": true, 00:13:54.117 "claim_type": "exclusive_write", 00:13:54.117 "zoned": false, 00:13:54.117 "supported_io_types": { 00:13:54.117 "read": true, 00:13:54.117 "write": true, 00:13:54.117 "unmap": true, 00:13:54.117 "flush": true, 00:13:54.117 "reset": true, 00:13:54.117 "nvme_admin": false, 00:13:54.117 "nvme_io": false, 00:13:54.117 "nvme_io_md": false, 00:13:54.117 "write_zeroes": true, 00:13:54.117 "zcopy": true, 00:13:54.117 "get_zone_info": false, 00:13:54.117 "zone_management": false, 00:13:54.117 "zone_append": false, 00:13:54.117 "compare": false, 00:13:54.117 "compare_and_write": false, 00:13:54.117 "abort": true, 00:13:54.117 "seek_hole": false, 00:13:54.117 "seek_data": false, 00:13:54.117 "copy": true, 00:13:54.117 "nvme_iov_md": false 00:13:54.117 }, 00:13:54.117 "memory_domains": [ 00:13:54.117 { 00:13:54.117 "dma_device_id": "system", 00:13:54.117 "dma_device_type": 1 00:13:54.117 }, 00:13:54.117 { 00:13:54.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:54.117 "dma_device_type": 2 00:13:54.117 } 00:13:54.117 ], 00:13:54.117 "driver_specific": {} 00:13:54.117 } 00:13:54.117 ] 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.117 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.118 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.118 "name": "Existed_Raid", 00:13:54.118 "uuid": "3704f5c8-d7c3-4528-81c6-5695ffa7cd86", 00:13:54.118 "strip_size_kb": 64, 00:13:54.118 "state": "online", 00:13:54.118 "raid_level": "raid5f", 00:13:54.118 "superblock": false, 00:13:54.118 "num_base_bdevs": 3, 00:13:54.118 "num_base_bdevs_discovered": 3, 00:13:54.118 "num_base_bdevs_operational": 3, 00:13:54.118 "base_bdevs_list": [ 00:13:54.118 { 00:13:54.118 "name": "NewBaseBdev", 00:13:54.118 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:54.118 "is_configured": true, 00:13:54.118 "data_offset": 0, 00:13:54.118 "data_size": 65536 00:13:54.118 }, 00:13:54.118 { 00:13:54.118 "name": "BaseBdev2", 00:13:54.118 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:54.118 "is_configured": true, 00:13:54.118 "data_offset": 0, 00:13:54.118 "data_size": 65536 00:13:54.118 }, 00:13:54.118 { 00:13:54.118 "name": "BaseBdev3", 00:13:54.118 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:54.118 "is_configured": true, 00:13:54.118 "data_offset": 0, 00:13:54.118 "data_size": 65536 00:13:54.118 } 00:13:54.118 ] 00:13:54.118 }' 00:13:54.118 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.118 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.396 [2024-10-29 11:02:59.799229] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:54.396 "name": "Existed_Raid", 00:13:54.396 "aliases": [ 00:13:54.396 "3704f5c8-d7c3-4528-81c6-5695ffa7cd86" 00:13:54.396 ], 00:13:54.396 "product_name": "Raid Volume", 00:13:54.396 "block_size": 512, 00:13:54.396 "num_blocks": 131072, 00:13:54.396 "uuid": "3704f5c8-d7c3-4528-81c6-5695ffa7cd86", 00:13:54.396 "assigned_rate_limits": { 00:13:54.396 "rw_ios_per_sec": 0, 00:13:54.396 "rw_mbytes_per_sec": 0, 00:13:54.396 "r_mbytes_per_sec": 0, 00:13:54.396 "w_mbytes_per_sec": 0 00:13:54.396 }, 00:13:54.396 "claimed": false, 00:13:54.396 "zoned": false, 00:13:54.396 "supported_io_types": { 00:13:54.396 "read": true, 00:13:54.396 "write": true, 00:13:54.396 "unmap": false, 00:13:54.396 "flush": false, 00:13:54.396 "reset": true, 00:13:54.396 "nvme_admin": false, 00:13:54.396 "nvme_io": false, 00:13:54.396 "nvme_io_md": false, 00:13:54.396 "write_zeroes": true, 00:13:54.396 "zcopy": false, 00:13:54.396 "get_zone_info": false, 00:13:54.396 "zone_management": false, 00:13:54.396 "zone_append": false, 00:13:54.396 "compare": false, 00:13:54.396 "compare_and_write": false, 00:13:54.396 "abort": false, 00:13:54.396 "seek_hole": false, 00:13:54.396 "seek_data": false, 00:13:54.396 "copy": false, 00:13:54.396 "nvme_iov_md": false 00:13:54.396 }, 00:13:54.396 "driver_specific": { 00:13:54.396 "raid": { 00:13:54.396 "uuid": "3704f5c8-d7c3-4528-81c6-5695ffa7cd86", 00:13:54.396 "strip_size_kb": 64, 00:13:54.396 "state": "online", 00:13:54.396 "raid_level": "raid5f", 00:13:54.396 "superblock": false, 00:13:54.396 "num_base_bdevs": 3, 00:13:54.396 "num_base_bdevs_discovered": 3, 00:13:54.396 "num_base_bdevs_operational": 3, 00:13:54.396 "base_bdevs_list": [ 00:13:54.396 { 00:13:54.396 "name": "NewBaseBdev", 00:13:54.396 "uuid": "3815fc25-b0b9-4fe1-8644-b2d017f9aa63", 00:13:54.396 "is_configured": true, 00:13:54.396 "data_offset": 0, 00:13:54.396 "data_size": 65536 00:13:54.396 }, 00:13:54.396 { 00:13:54.396 "name": "BaseBdev2", 00:13:54.396 "uuid": "626e691a-e973-41a6-9904-111e9d200293", 00:13:54.396 "is_configured": true, 00:13:54.396 "data_offset": 0, 00:13:54.396 "data_size": 65536 00:13:54.396 }, 00:13:54.396 { 00:13:54.396 "name": "BaseBdev3", 00:13:54.396 "uuid": "9bd48a7f-394a-48fe-ac8e-ae1b4e51628e", 00:13:54.396 "is_configured": true, 00:13:54.396 "data_offset": 0, 00:13:54.396 "data_size": 65536 00:13:54.396 } 00:13:54.396 ] 00:13:54.396 } 00:13:54.396 } 00:13:54.396 }' 00:13:54.396 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:54.680 BaseBdev2 00:13:54.680 BaseBdev3' 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.680 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.681 11:02:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.681 [2024-10-29 11:03:00.094516] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:54.681 [2024-10-29 11:03:00.094590] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:54.681 [2024-10-29 11:03:00.094682] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:54.681 [2024-10-29 11:03:00.094982] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:54.681 [2024-10-29 11:03:00.095047] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90566 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 90566 ']' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # kill -0 90566 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@957 -- # uname 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 90566 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:13:54.681 killing process with pid 90566 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 90566' 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@971 -- # kill 90566 00:13:54.681 [2024-10-29 11:03:00.133223] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:54.681 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@976 -- # wait 90566 00:13:54.681 [2024-10-29 11:03:00.164166] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:54.948 11:03:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:54.948 ************************************ 00:13:54.948 END TEST raid5f_state_function_test 00:13:54.948 ************************************ 00:13:54.948 00:13:54.948 real 0m8.902s 00:13:54.948 user 0m15.166s 00:13:54.948 sys 0m1.872s 00:13:54.948 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:13:54.948 11:03:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.948 11:03:00 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:54.948 11:03:00 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:13:54.948 11:03:00 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:13:54.948 11:03:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:55.207 ************************************ 00:13:55.207 START TEST raid5f_state_function_test_sb 00:13:55.207 ************************************ 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid5f 3 true 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:55.207 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=91165 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 91165' 00:13:55.208 Process raid pid: 91165 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 91165 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 91165 ']' 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:55.208 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:13:55.208 11:03:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:55.208 [2024-10-29 11:03:00.558927] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:13:55.208 [2024-10-29 11:03:00.559074] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:55.468 [2024-10-29 11:03:00.709828] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:55.468 [2024-10-29 11:03:00.735403] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:55.468 [2024-10-29 11:03:00.779381] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:55.468 [2024-10-29 11:03:00.779515] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.038 [2024-10-29 11:03:01.385785] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:56.038 [2024-10-29 11:03:01.385854] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:56.038 [2024-10-29 11:03:01.385882] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:56.038 [2024-10-29 11:03:01.385896] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:56.038 [2024-10-29 11:03:01.385907] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:56.038 [2024-10-29 11:03:01.385921] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.038 "name": "Existed_Raid", 00:13:56.038 "uuid": "f7aa4ab6-bba5-4157-9764-8a025332b4bc", 00:13:56.038 "strip_size_kb": 64, 00:13:56.038 "state": "configuring", 00:13:56.038 "raid_level": "raid5f", 00:13:56.038 "superblock": true, 00:13:56.038 "num_base_bdevs": 3, 00:13:56.038 "num_base_bdevs_discovered": 0, 00:13:56.038 "num_base_bdevs_operational": 3, 00:13:56.038 "base_bdevs_list": [ 00:13:56.038 { 00:13:56.038 "name": "BaseBdev1", 00:13:56.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.038 "is_configured": false, 00:13:56.038 "data_offset": 0, 00:13:56.038 "data_size": 0 00:13:56.038 }, 00:13:56.038 { 00:13:56.038 "name": "BaseBdev2", 00:13:56.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.038 "is_configured": false, 00:13:56.038 "data_offset": 0, 00:13:56.038 "data_size": 0 00:13:56.038 }, 00:13:56.038 { 00:13:56.038 "name": "BaseBdev3", 00:13:56.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.038 "is_configured": false, 00:13:56.038 "data_offset": 0, 00:13:56.038 "data_size": 0 00:13:56.038 } 00:13:56.038 ] 00:13:56.038 }' 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.038 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.299 [2024-10-29 11:03:01.793003] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:56.299 [2024-10-29 11:03:01.793099] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.299 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.559 [2024-10-29 11:03:01.801024] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:56.559 [2024-10-29 11:03:01.801114] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:56.559 [2024-10-29 11:03:01.801161] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:56.559 [2024-10-29 11:03:01.801188] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:56.559 [2024-10-29 11:03:01.801210] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:56.559 [2024-10-29 11:03:01.801236] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.559 [2024-10-29 11:03:01.818069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:56.559 BaseBdev1 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.559 [ 00:13:56.559 { 00:13:56.559 "name": "BaseBdev1", 00:13:56.559 "aliases": [ 00:13:56.559 "87dc45f8-0656-4c9d-9bfc-69fa585638f6" 00:13:56.559 ], 00:13:56.559 "product_name": "Malloc disk", 00:13:56.559 "block_size": 512, 00:13:56.559 "num_blocks": 65536, 00:13:56.559 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:56.559 "assigned_rate_limits": { 00:13:56.559 "rw_ios_per_sec": 0, 00:13:56.559 "rw_mbytes_per_sec": 0, 00:13:56.559 "r_mbytes_per_sec": 0, 00:13:56.559 "w_mbytes_per_sec": 0 00:13:56.559 }, 00:13:56.559 "claimed": true, 00:13:56.559 "claim_type": "exclusive_write", 00:13:56.559 "zoned": false, 00:13:56.559 "supported_io_types": { 00:13:56.559 "read": true, 00:13:56.559 "write": true, 00:13:56.559 "unmap": true, 00:13:56.559 "flush": true, 00:13:56.559 "reset": true, 00:13:56.559 "nvme_admin": false, 00:13:56.559 "nvme_io": false, 00:13:56.559 "nvme_io_md": false, 00:13:56.559 "write_zeroes": true, 00:13:56.559 "zcopy": true, 00:13:56.559 "get_zone_info": false, 00:13:56.559 "zone_management": false, 00:13:56.559 "zone_append": false, 00:13:56.559 "compare": false, 00:13:56.559 "compare_and_write": false, 00:13:56.559 "abort": true, 00:13:56.559 "seek_hole": false, 00:13:56.559 "seek_data": false, 00:13:56.559 "copy": true, 00:13:56.559 "nvme_iov_md": false 00:13:56.559 }, 00:13:56.559 "memory_domains": [ 00:13:56.559 { 00:13:56.559 "dma_device_id": "system", 00:13:56.559 "dma_device_type": 1 00:13:56.559 }, 00:13:56.559 { 00:13:56.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:56.559 "dma_device_type": 2 00:13:56.559 } 00:13:56.559 ], 00:13:56.559 "driver_specific": {} 00:13:56.559 } 00:13:56.559 ] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.559 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.560 "name": "Existed_Raid", 00:13:56.560 "uuid": "75dc3581-b18b-43c5-acaa-5e5755c35fe9", 00:13:56.560 "strip_size_kb": 64, 00:13:56.560 "state": "configuring", 00:13:56.560 "raid_level": "raid5f", 00:13:56.560 "superblock": true, 00:13:56.560 "num_base_bdevs": 3, 00:13:56.560 "num_base_bdevs_discovered": 1, 00:13:56.560 "num_base_bdevs_operational": 3, 00:13:56.560 "base_bdevs_list": [ 00:13:56.560 { 00:13:56.560 "name": "BaseBdev1", 00:13:56.560 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:56.560 "is_configured": true, 00:13:56.560 "data_offset": 2048, 00:13:56.560 "data_size": 63488 00:13:56.560 }, 00:13:56.560 { 00:13:56.560 "name": "BaseBdev2", 00:13:56.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.560 "is_configured": false, 00:13:56.560 "data_offset": 0, 00:13:56.560 "data_size": 0 00:13:56.560 }, 00:13:56.560 { 00:13:56.560 "name": "BaseBdev3", 00:13:56.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.560 "is_configured": false, 00:13:56.560 "data_offset": 0, 00:13:56.560 "data_size": 0 00:13:56.560 } 00:13:56.560 ] 00:13:56.560 }' 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.560 11:03:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.820 [2024-10-29 11:03:02.229429] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:56.820 [2024-10-29 11:03:02.229525] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.820 [2024-10-29 11:03:02.241446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:56.820 [2024-10-29 11:03:02.243314] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:56.820 [2024-10-29 11:03:02.243427] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:56.820 [2024-10-29 11:03:02.243466] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:56.820 [2024-10-29 11:03:02.243495] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.820 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.821 "name": "Existed_Raid", 00:13:56.821 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:56.821 "strip_size_kb": 64, 00:13:56.821 "state": "configuring", 00:13:56.821 "raid_level": "raid5f", 00:13:56.821 "superblock": true, 00:13:56.821 "num_base_bdevs": 3, 00:13:56.821 "num_base_bdevs_discovered": 1, 00:13:56.821 "num_base_bdevs_operational": 3, 00:13:56.821 "base_bdevs_list": [ 00:13:56.821 { 00:13:56.821 "name": "BaseBdev1", 00:13:56.821 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:56.821 "is_configured": true, 00:13:56.821 "data_offset": 2048, 00:13:56.821 "data_size": 63488 00:13:56.821 }, 00:13:56.821 { 00:13:56.821 "name": "BaseBdev2", 00:13:56.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.821 "is_configured": false, 00:13:56.821 "data_offset": 0, 00:13:56.821 "data_size": 0 00:13:56.821 }, 00:13:56.821 { 00:13:56.821 "name": "BaseBdev3", 00:13:56.821 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.821 "is_configured": false, 00:13:56.821 "data_offset": 0, 00:13:56.821 "data_size": 0 00:13:56.821 } 00:13:56.821 ] 00:13:56.821 }' 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.821 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 [2024-10-29 11:03:02.667896] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:57.392 BaseBdev2 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 [ 00:13:57.392 { 00:13:57.392 "name": "BaseBdev2", 00:13:57.392 "aliases": [ 00:13:57.392 "8ab50323-5400-48fb-b17c-c0365b9d497a" 00:13:57.392 ], 00:13:57.392 "product_name": "Malloc disk", 00:13:57.392 "block_size": 512, 00:13:57.392 "num_blocks": 65536, 00:13:57.392 "uuid": "8ab50323-5400-48fb-b17c-c0365b9d497a", 00:13:57.392 "assigned_rate_limits": { 00:13:57.392 "rw_ios_per_sec": 0, 00:13:57.392 "rw_mbytes_per_sec": 0, 00:13:57.392 "r_mbytes_per_sec": 0, 00:13:57.392 "w_mbytes_per_sec": 0 00:13:57.392 }, 00:13:57.392 "claimed": true, 00:13:57.392 "claim_type": "exclusive_write", 00:13:57.392 "zoned": false, 00:13:57.392 "supported_io_types": { 00:13:57.392 "read": true, 00:13:57.392 "write": true, 00:13:57.392 "unmap": true, 00:13:57.392 "flush": true, 00:13:57.392 "reset": true, 00:13:57.392 "nvme_admin": false, 00:13:57.392 "nvme_io": false, 00:13:57.392 "nvme_io_md": false, 00:13:57.392 "write_zeroes": true, 00:13:57.392 "zcopy": true, 00:13:57.392 "get_zone_info": false, 00:13:57.392 "zone_management": false, 00:13:57.392 "zone_append": false, 00:13:57.392 "compare": false, 00:13:57.392 "compare_and_write": false, 00:13:57.392 "abort": true, 00:13:57.392 "seek_hole": false, 00:13:57.392 "seek_data": false, 00:13:57.392 "copy": true, 00:13:57.392 "nvme_iov_md": false 00:13:57.392 }, 00:13:57.392 "memory_domains": [ 00:13:57.392 { 00:13:57.392 "dma_device_id": "system", 00:13:57.392 "dma_device_type": 1 00:13:57.392 }, 00:13:57.392 { 00:13:57.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:57.392 "dma_device_type": 2 00:13:57.392 } 00:13:57.392 ], 00:13:57.392 "driver_specific": {} 00:13:57.392 } 00:13:57.392 ] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.392 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.392 "name": "Existed_Raid", 00:13:57.392 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:57.392 "strip_size_kb": 64, 00:13:57.392 "state": "configuring", 00:13:57.392 "raid_level": "raid5f", 00:13:57.392 "superblock": true, 00:13:57.392 "num_base_bdevs": 3, 00:13:57.392 "num_base_bdevs_discovered": 2, 00:13:57.392 "num_base_bdevs_operational": 3, 00:13:57.392 "base_bdevs_list": [ 00:13:57.392 { 00:13:57.392 "name": "BaseBdev1", 00:13:57.392 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:57.392 "is_configured": true, 00:13:57.392 "data_offset": 2048, 00:13:57.392 "data_size": 63488 00:13:57.392 }, 00:13:57.392 { 00:13:57.392 "name": "BaseBdev2", 00:13:57.392 "uuid": "8ab50323-5400-48fb-b17c-c0365b9d497a", 00:13:57.392 "is_configured": true, 00:13:57.392 "data_offset": 2048, 00:13:57.392 "data_size": 63488 00:13:57.392 }, 00:13:57.392 { 00:13:57.392 "name": "BaseBdev3", 00:13:57.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:57.393 "is_configured": false, 00:13:57.393 "data_offset": 0, 00:13:57.393 "data_size": 0 00:13:57.393 } 00:13:57.393 ] 00:13:57.393 }' 00:13:57.393 11:03:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.393 11:03:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.653 [2024-10-29 11:03:03.146811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:57.653 [2024-10-29 11:03:03.147829] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:13:57.653 [2024-10-29 11:03:03.147916] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:57.653 BaseBdev3 00:13:57.653 [2024-10-29 11:03:03.149009] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:57.653 [2024-10-29 11:03:03.150748] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:13:57.653 [2024-10-29 11:03:03.150821] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.653 [2024-10-29 11:03:03.151358] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.653 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 [ 00:13:57.914 { 00:13:57.914 "name": "BaseBdev3", 00:13:57.914 "aliases": [ 00:13:57.914 "085b1579-55a0-4752-9495-22fde1ac87eb" 00:13:57.914 ], 00:13:57.914 "product_name": "Malloc disk", 00:13:57.914 "block_size": 512, 00:13:57.914 "num_blocks": 65536, 00:13:57.914 "uuid": "085b1579-55a0-4752-9495-22fde1ac87eb", 00:13:57.914 "assigned_rate_limits": { 00:13:57.914 "rw_ios_per_sec": 0, 00:13:57.914 "rw_mbytes_per_sec": 0, 00:13:57.914 "r_mbytes_per_sec": 0, 00:13:57.914 "w_mbytes_per_sec": 0 00:13:57.914 }, 00:13:57.914 "claimed": true, 00:13:57.914 "claim_type": "exclusive_write", 00:13:57.914 "zoned": false, 00:13:57.914 "supported_io_types": { 00:13:57.914 "read": true, 00:13:57.914 "write": true, 00:13:57.914 "unmap": true, 00:13:57.914 "flush": true, 00:13:57.914 "reset": true, 00:13:57.914 "nvme_admin": false, 00:13:57.914 "nvme_io": false, 00:13:57.914 "nvme_io_md": false, 00:13:57.914 "write_zeroes": true, 00:13:57.914 "zcopy": true, 00:13:57.914 "get_zone_info": false, 00:13:57.914 "zone_management": false, 00:13:57.914 "zone_append": false, 00:13:57.914 "compare": false, 00:13:57.914 "compare_and_write": false, 00:13:57.914 "abort": true, 00:13:57.914 "seek_hole": false, 00:13:57.914 "seek_data": false, 00:13:57.914 "copy": true, 00:13:57.914 "nvme_iov_md": false 00:13:57.914 }, 00:13:57.914 "memory_domains": [ 00:13:57.914 { 00:13:57.914 "dma_device_id": "system", 00:13:57.914 "dma_device_type": 1 00:13:57.914 }, 00:13:57.914 { 00:13:57.914 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:57.914 "dma_device_type": 2 00:13:57.914 } 00:13:57.914 ], 00:13:57.914 "driver_specific": {} 00:13:57.914 } 00:13:57.914 ] 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.914 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.914 "name": "Existed_Raid", 00:13:57.914 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:57.914 "strip_size_kb": 64, 00:13:57.914 "state": "online", 00:13:57.914 "raid_level": "raid5f", 00:13:57.914 "superblock": true, 00:13:57.914 "num_base_bdevs": 3, 00:13:57.914 "num_base_bdevs_discovered": 3, 00:13:57.915 "num_base_bdevs_operational": 3, 00:13:57.915 "base_bdevs_list": [ 00:13:57.915 { 00:13:57.915 "name": "BaseBdev1", 00:13:57.915 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 }, 00:13:57.915 { 00:13:57.915 "name": "BaseBdev2", 00:13:57.915 "uuid": "8ab50323-5400-48fb-b17c-c0365b9d497a", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 }, 00:13:57.915 { 00:13:57.915 "name": "BaseBdev3", 00:13:57.915 "uuid": "085b1579-55a0-4752-9495-22fde1ac87eb", 00:13:57.915 "is_configured": true, 00:13:57.915 "data_offset": 2048, 00:13:57.915 "data_size": 63488 00:13:57.915 } 00:13:57.915 ] 00:13:57.915 }' 00:13:57.915 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.915 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.176 [2024-10-29 11:03:03.590459] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:58.176 "name": "Existed_Raid", 00:13:58.176 "aliases": [ 00:13:58.176 "be1548f6-378f-43b2-bb35-e6e4bc6b46be" 00:13:58.176 ], 00:13:58.176 "product_name": "Raid Volume", 00:13:58.176 "block_size": 512, 00:13:58.176 "num_blocks": 126976, 00:13:58.176 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:58.176 "assigned_rate_limits": { 00:13:58.176 "rw_ios_per_sec": 0, 00:13:58.176 "rw_mbytes_per_sec": 0, 00:13:58.176 "r_mbytes_per_sec": 0, 00:13:58.176 "w_mbytes_per_sec": 0 00:13:58.176 }, 00:13:58.176 "claimed": false, 00:13:58.176 "zoned": false, 00:13:58.176 "supported_io_types": { 00:13:58.176 "read": true, 00:13:58.176 "write": true, 00:13:58.176 "unmap": false, 00:13:58.176 "flush": false, 00:13:58.176 "reset": true, 00:13:58.176 "nvme_admin": false, 00:13:58.176 "nvme_io": false, 00:13:58.176 "nvme_io_md": false, 00:13:58.176 "write_zeroes": true, 00:13:58.176 "zcopy": false, 00:13:58.176 "get_zone_info": false, 00:13:58.176 "zone_management": false, 00:13:58.176 "zone_append": false, 00:13:58.176 "compare": false, 00:13:58.176 "compare_and_write": false, 00:13:58.176 "abort": false, 00:13:58.176 "seek_hole": false, 00:13:58.176 "seek_data": false, 00:13:58.176 "copy": false, 00:13:58.176 "nvme_iov_md": false 00:13:58.176 }, 00:13:58.176 "driver_specific": { 00:13:58.176 "raid": { 00:13:58.176 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:58.176 "strip_size_kb": 64, 00:13:58.176 "state": "online", 00:13:58.176 "raid_level": "raid5f", 00:13:58.176 "superblock": true, 00:13:58.176 "num_base_bdevs": 3, 00:13:58.176 "num_base_bdevs_discovered": 3, 00:13:58.176 "num_base_bdevs_operational": 3, 00:13:58.176 "base_bdevs_list": [ 00:13:58.176 { 00:13:58.176 "name": "BaseBdev1", 00:13:58.176 "uuid": "87dc45f8-0656-4c9d-9bfc-69fa585638f6", 00:13:58.176 "is_configured": true, 00:13:58.176 "data_offset": 2048, 00:13:58.176 "data_size": 63488 00:13:58.176 }, 00:13:58.176 { 00:13:58.176 "name": "BaseBdev2", 00:13:58.176 "uuid": "8ab50323-5400-48fb-b17c-c0365b9d497a", 00:13:58.176 "is_configured": true, 00:13:58.176 "data_offset": 2048, 00:13:58.176 "data_size": 63488 00:13:58.176 }, 00:13:58.176 { 00:13:58.176 "name": "BaseBdev3", 00:13:58.176 "uuid": "085b1579-55a0-4752-9495-22fde1ac87eb", 00:13:58.176 "is_configured": true, 00:13:58.176 "data_offset": 2048, 00:13:58.176 "data_size": 63488 00:13:58.176 } 00:13:58.176 ] 00:13:58.176 } 00:13:58.176 } 00:13:58.176 }' 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:58.176 BaseBdev2 00:13:58.176 BaseBdev3' 00:13:58.176 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.438 [2024-10-29 11:03:03.837841] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:58.438 "name": "Existed_Raid", 00:13:58.438 "uuid": "be1548f6-378f-43b2-bb35-e6e4bc6b46be", 00:13:58.438 "strip_size_kb": 64, 00:13:58.438 "state": "online", 00:13:58.438 "raid_level": "raid5f", 00:13:58.438 "superblock": true, 00:13:58.438 "num_base_bdevs": 3, 00:13:58.438 "num_base_bdevs_discovered": 2, 00:13:58.438 "num_base_bdevs_operational": 2, 00:13:58.438 "base_bdevs_list": [ 00:13:58.438 { 00:13:58.438 "name": null, 00:13:58.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:58.438 "is_configured": false, 00:13:58.438 "data_offset": 0, 00:13:58.438 "data_size": 63488 00:13:58.438 }, 00:13:58.438 { 00:13:58.438 "name": "BaseBdev2", 00:13:58.438 "uuid": "8ab50323-5400-48fb-b17c-c0365b9d497a", 00:13:58.438 "is_configured": true, 00:13:58.438 "data_offset": 2048, 00:13:58.438 "data_size": 63488 00:13:58.438 }, 00:13:58.438 { 00:13:58.438 "name": "BaseBdev3", 00:13:58.438 "uuid": "085b1579-55a0-4752-9495-22fde1ac87eb", 00:13:58.438 "is_configured": true, 00:13:58.438 "data_offset": 2048, 00:13:58.438 "data_size": 63488 00:13:58.438 } 00:13:58.438 ] 00:13:58.438 }' 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:58.438 11:03:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 [2024-10-29 11:03:04.260631] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:59.008 [2024-10-29 11:03:04.260841] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:59.008 [2024-10-29 11:03:04.272056] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 [2024-10-29 11:03:04.331976] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:59.008 [2024-10-29 11:03:04.332025] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:59.008 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 BaseBdev2 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 [ 00:13:59.009 { 00:13:59.009 "name": "BaseBdev2", 00:13:59.009 "aliases": [ 00:13:59.009 "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c" 00:13:59.009 ], 00:13:59.009 "product_name": "Malloc disk", 00:13:59.009 "block_size": 512, 00:13:59.009 "num_blocks": 65536, 00:13:59.009 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:13:59.009 "assigned_rate_limits": { 00:13:59.009 "rw_ios_per_sec": 0, 00:13:59.009 "rw_mbytes_per_sec": 0, 00:13:59.009 "r_mbytes_per_sec": 0, 00:13:59.009 "w_mbytes_per_sec": 0 00:13:59.009 }, 00:13:59.009 "claimed": false, 00:13:59.009 "zoned": false, 00:13:59.009 "supported_io_types": { 00:13:59.009 "read": true, 00:13:59.009 "write": true, 00:13:59.009 "unmap": true, 00:13:59.009 "flush": true, 00:13:59.009 "reset": true, 00:13:59.009 "nvme_admin": false, 00:13:59.009 "nvme_io": false, 00:13:59.009 "nvme_io_md": false, 00:13:59.009 "write_zeroes": true, 00:13:59.009 "zcopy": true, 00:13:59.009 "get_zone_info": false, 00:13:59.009 "zone_management": false, 00:13:59.009 "zone_append": false, 00:13:59.009 "compare": false, 00:13:59.009 "compare_and_write": false, 00:13:59.009 "abort": true, 00:13:59.009 "seek_hole": false, 00:13:59.009 "seek_data": false, 00:13:59.009 "copy": true, 00:13:59.009 "nvme_iov_md": false 00:13:59.009 }, 00:13:59.009 "memory_domains": [ 00:13:59.009 { 00:13:59.009 "dma_device_id": "system", 00:13:59.009 "dma_device_type": 1 00:13:59.009 }, 00:13:59.009 { 00:13:59.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:59.009 "dma_device_type": 2 00:13:59.009 } 00:13:59.009 ], 00:13:59.009 "driver_specific": {} 00:13:59.009 } 00:13:59.009 ] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 BaseBdev3 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.009 [ 00:13:59.009 { 00:13:59.009 "name": "BaseBdev3", 00:13:59.009 "aliases": [ 00:13:59.009 "ac0a43e2-e204-45de-b0dc-73a11a91de6a" 00:13:59.009 ], 00:13:59.009 "product_name": "Malloc disk", 00:13:59.009 "block_size": 512, 00:13:59.009 "num_blocks": 65536, 00:13:59.009 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:13:59.009 "assigned_rate_limits": { 00:13:59.009 "rw_ios_per_sec": 0, 00:13:59.009 "rw_mbytes_per_sec": 0, 00:13:59.009 "r_mbytes_per_sec": 0, 00:13:59.009 "w_mbytes_per_sec": 0 00:13:59.009 }, 00:13:59.009 "claimed": false, 00:13:59.009 "zoned": false, 00:13:59.009 "supported_io_types": { 00:13:59.009 "read": true, 00:13:59.009 "write": true, 00:13:59.009 "unmap": true, 00:13:59.009 "flush": true, 00:13:59.009 "reset": true, 00:13:59.009 "nvme_admin": false, 00:13:59.009 "nvme_io": false, 00:13:59.009 "nvme_io_md": false, 00:13:59.009 "write_zeroes": true, 00:13:59.009 "zcopy": true, 00:13:59.009 "get_zone_info": false, 00:13:59.009 "zone_management": false, 00:13:59.009 "zone_append": false, 00:13:59.009 "compare": false, 00:13:59.009 "compare_and_write": false, 00:13:59.009 "abort": true, 00:13:59.009 "seek_hole": false, 00:13:59.009 "seek_data": false, 00:13:59.009 "copy": true, 00:13:59.009 "nvme_iov_md": false 00:13:59.009 }, 00:13:59.009 "memory_domains": [ 00:13:59.009 { 00:13:59.009 "dma_device_id": "system", 00:13:59.009 "dma_device_type": 1 00:13:59.009 }, 00:13:59.009 { 00:13:59.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:59.009 "dma_device_type": 2 00:13:59.009 } 00:13:59.009 ], 00:13:59.009 "driver_specific": {} 00:13:59.009 } 00:13:59.009 ] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.009 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.269 [2024-10-29 11:03:04.507699] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:59.269 [2024-10-29 11:03:04.507813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:59.269 [2024-10-29 11:03:04.507877] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:59.269 [2024-10-29 11:03:04.509690] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.269 "name": "Existed_Raid", 00:13:59.269 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:13:59.269 "strip_size_kb": 64, 00:13:59.269 "state": "configuring", 00:13:59.269 "raid_level": "raid5f", 00:13:59.269 "superblock": true, 00:13:59.269 "num_base_bdevs": 3, 00:13:59.269 "num_base_bdevs_discovered": 2, 00:13:59.269 "num_base_bdevs_operational": 3, 00:13:59.269 "base_bdevs_list": [ 00:13:59.269 { 00:13:59.269 "name": "BaseBdev1", 00:13:59.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.269 "is_configured": false, 00:13:59.269 "data_offset": 0, 00:13:59.269 "data_size": 0 00:13:59.269 }, 00:13:59.269 { 00:13:59.269 "name": "BaseBdev2", 00:13:59.269 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:13:59.269 "is_configured": true, 00:13:59.269 "data_offset": 2048, 00:13:59.269 "data_size": 63488 00:13:59.269 }, 00:13:59.269 { 00:13:59.269 "name": "BaseBdev3", 00:13:59.269 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:13:59.269 "is_configured": true, 00:13:59.269 "data_offset": 2048, 00:13:59.269 "data_size": 63488 00:13:59.269 } 00:13:59.269 ] 00:13:59.269 }' 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.269 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.528 [2024-10-29 11:03:04.930933] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.528 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.528 "name": "Existed_Raid", 00:13:59.528 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:13:59.528 "strip_size_kb": 64, 00:13:59.528 "state": "configuring", 00:13:59.528 "raid_level": "raid5f", 00:13:59.528 "superblock": true, 00:13:59.528 "num_base_bdevs": 3, 00:13:59.528 "num_base_bdevs_discovered": 1, 00:13:59.528 "num_base_bdevs_operational": 3, 00:13:59.528 "base_bdevs_list": [ 00:13:59.528 { 00:13:59.528 "name": "BaseBdev1", 00:13:59.528 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.529 "is_configured": false, 00:13:59.529 "data_offset": 0, 00:13:59.529 "data_size": 0 00:13:59.529 }, 00:13:59.529 { 00:13:59.529 "name": null, 00:13:59.529 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:13:59.529 "is_configured": false, 00:13:59.529 "data_offset": 0, 00:13:59.529 "data_size": 63488 00:13:59.529 }, 00:13:59.529 { 00:13:59.529 "name": "BaseBdev3", 00:13:59.529 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:13:59.529 "is_configured": true, 00:13:59.529 "data_offset": 2048, 00:13:59.529 "data_size": 63488 00:13:59.529 } 00:13:59.529 ] 00:13:59.529 }' 00:13:59.529 11:03:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.529 11:03:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 [2024-10-29 11:03:05.453290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:00.098 BaseBdev1 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 [ 00:14:00.098 { 00:14:00.098 "name": "BaseBdev1", 00:14:00.098 "aliases": [ 00:14:00.098 "0b5a2d29-feb2-4550-a784-854322ece7f7" 00:14:00.098 ], 00:14:00.098 "product_name": "Malloc disk", 00:14:00.098 "block_size": 512, 00:14:00.098 "num_blocks": 65536, 00:14:00.098 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:00.098 "assigned_rate_limits": { 00:14:00.098 "rw_ios_per_sec": 0, 00:14:00.098 "rw_mbytes_per_sec": 0, 00:14:00.098 "r_mbytes_per_sec": 0, 00:14:00.098 "w_mbytes_per_sec": 0 00:14:00.098 }, 00:14:00.098 "claimed": true, 00:14:00.098 "claim_type": "exclusive_write", 00:14:00.098 "zoned": false, 00:14:00.098 "supported_io_types": { 00:14:00.098 "read": true, 00:14:00.098 "write": true, 00:14:00.098 "unmap": true, 00:14:00.098 "flush": true, 00:14:00.098 "reset": true, 00:14:00.098 "nvme_admin": false, 00:14:00.098 "nvme_io": false, 00:14:00.098 "nvme_io_md": false, 00:14:00.098 "write_zeroes": true, 00:14:00.098 "zcopy": true, 00:14:00.098 "get_zone_info": false, 00:14:00.098 "zone_management": false, 00:14:00.098 "zone_append": false, 00:14:00.098 "compare": false, 00:14:00.098 "compare_and_write": false, 00:14:00.098 "abort": true, 00:14:00.098 "seek_hole": false, 00:14:00.098 "seek_data": false, 00:14:00.098 "copy": true, 00:14:00.098 "nvme_iov_md": false 00:14:00.098 }, 00:14:00.098 "memory_domains": [ 00:14:00.098 { 00:14:00.098 "dma_device_id": "system", 00:14:00.098 "dma_device_type": 1 00:14:00.098 }, 00:14:00.098 { 00:14:00.098 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:00.098 "dma_device_type": 2 00:14:00.098 } 00:14:00.098 ], 00:14:00.098 "driver_specific": {} 00:14:00.098 } 00:14:00.098 ] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.098 "name": "Existed_Raid", 00:14:00.098 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:00.098 "strip_size_kb": 64, 00:14:00.098 "state": "configuring", 00:14:00.098 "raid_level": "raid5f", 00:14:00.098 "superblock": true, 00:14:00.098 "num_base_bdevs": 3, 00:14:00.098 "num_base_bdevs_discovered": 2, 00:14:00.098 "num_base_bdevs_operational": 3, 00:14:00.098 "base_bdevs_list": [ 00:14:00.098 { 00:14:00.098 "name": "BaseBdev1", 00:14:00.098 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:00.098 "is_configured": true, 00:14:00.098 "data_offset": 2048, 00:14:00.098 "data_size": 63488 00:14:00.098 }, 00:14:00.098 { 00:14:00.098 "name": null, 00:14:00.098 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:00.098 "is_configured": false, 00:14:00.098 "data_offset": 0, 00:14:00.098 "data_size": 63488 00:14:00.098 }, 00:14:00.098 { 00:14:00.098 "name": "BaseBdev3", 00:14:00.098 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:00.098 "is_configured": true, 00:14:00.098 "data_offset": 2048, 00:14:00.098 "data_size": 63488 00:14:00.098 } 00:14:00.098 ] 00:14:00.098 }' 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.098 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.668 [2024-10-29 11:03:05.972478] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:00.668 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.669 11:03:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.669 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.669 "name": "Existed_Raid", 00:14:00.669 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:00.669 "strip_size_kb": 64, 00:14:00.669 "state": "configuring", 00:14:00.669 "raid_level": "raid5f", 00:14:00.669 "superblock": true, 00:14:00.669 "num_base_bdevs": 3, 00:14:00.669 "num_base_bdevs_discovered": 1, 00:14:00.669 "num_base_bdevs_operational": 3, 00:14:00.669 "base_bdevs_list": [ 00:14:00.669 { 00:14:00.669 "name": "BaseBdev1", 00:14:00.669 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:00.669 "is_configured": true, 00:14:00.669 "data_offset": 2048, 00:14:00.669 "data_size": 63488 00:14:00.669 }, 00:14:00.669 { 00:14:00.669 "name": null, 00:14:00.669 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:00.669 "is_configured": false, 00:14:00.669 "data_offset": 0, 00:14:00.669 "data_size": 63488 00:14:00.669 }, 00:14:00.669 { 00:14:00.669 "name": null, 00:14:00.669 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:00.669 "is_configured": false, 00:14:00.669 "data_offset": 0, 00:14:00.669 "data_size": 63488 00:14:00.669 } 00:14:00.669 ] 00:14:00.669 }' 00:14:00.669 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.669 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.928 [2024-10-29 11:03:06.407707] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.928 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.188 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.188 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.188 "name": "Existed_Raid", 00:14:01.188 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:01.188 "strip_size_kb": 64, 00:14:01.188 "state": "configuring", 00:14:01.188 "raid_level": "raid5f", 00:14:01.188 "superblock": true, 00:14:01.188 "num_base_bdevs": 3, 00:14:01.188 "num_base_bdevs_discovered": 2, 00:14:01.188 "num_base_bdevs_operational": 3, 00:14:01.188 "base_bdevs_list": [ 00:14:01.188 { 00:14:01.188 "name": "BaseBdev1", 00:14:01.188 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:01.188 "is_configured": true, 00:14:01.188 "data_offset": 2048, 00:14:01.188 "data_size": 63488 00:14:01.188 }, 00:14:01.188 { 00:14:01.188 "name": null, 00:14:01.188 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:01.188 "is_configured": false, 00:14:01.188 "data_offset": 0, 00:14:01.188 "data_size": 63488 00:14:01.188 }, 00:14:01.188 { 00:14:01.188 "name": "BaseBdev3", 00:14:01.189 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:01.189 "is_configured": true, 00:14:01.189 "data_offset": 2048, 00:14:01.189 "data_size": 63488 00:14:01.189 } 00:14:01.189 ] 00:14:01.189 }' 00:14:01.189 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.189 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.449 [2024-10-29 11:03:06.855001] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.449 "name": "Existed_Raid", 00:14:01.449 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:01.449 "strip_size_kb": 64, 00:14:01.449 "state": "configuring", 00:14:01.449 "raid_level": "raid5f", 00:14:01.449 "superblock": true, 00:14:01.449 "num_base_bdevs": 3, 00:14:01.449 "num_base_bdevs_discovered": 1, 00:14:01.449 "num_base_bdevs_operational": 3, 00:14:01.449 "base_bdevs_list": [ 00:14:01.449 { 00:14:01.449 "name": null, 00:14:01.449 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:01.449 "is_configured": false, 00:14:01.449 "data_offset": 0, 00:14:01.449 "data_size": 63488 00:14:01.449 }, 00:14:01.449 { 00:14:01.449 "name": null, 00:14:01.449 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:01.449 "is_configured": false, 00:14:01.449 "data_offset": 0, 00:14:01.449 "data_size": 63488 00:14:01.449 }, 00:14:01.449 { 00:14:01.449 "name": "BaseBdev3", 00:14:01.449 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:01.449 "is_configured": true, 00:14:01.449 "data_offset": 2048, 00:14:01.449 "data_size": 63488 00:14:01.449 } 00:14:01.449 ] 00:14:01.449 }' 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.449 11:03:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.020 [2024-10-29 11:03:07.332855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.020 "name": "Existed_Raid", 00:14:02.020 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:02.020 "strip_size_kb": 64, 00:14:02.020 "state": "configuring", 00:14:02.020 "raid_level": "raid5f", 00:14:02.020 "superblock": true, 00:14:02.020 "num_base_bdevs": 3, 00:14:02.020 "num_base_bdevs_discovered": 2, 00:14:02.020 "num_base_bdevs_operational": 3, 00:14:02.020 "base_bdevs_list": [ 00:14:02.020 { 00:14:02.020 "name": null, 00:14:02.020 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:02.020 "is_configured": false, 00:14:02.020 "data_offset": 0, 00:14:02.020 "data_size": 63488 00:14:02.020 }, 00:14:02.020 { 00:14:02.020 "name": "BaseBdev2", 00:14:02.020 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:02.020 "is_configured": true, 00:14:02.020 "data_offset": 2048, 00:14:02.020 "data_size": 63488 00:14:02.020 }, 00:14:02.020 { 00:14:02.020 "name": "BaseBdev3", 00:14:02.020 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:02.020 "is_configured": true, 00:14:02.020 "data_offset": 2048, 00:14:02.020 "data_size": 63488 00:14:02.020 } 00:14:02.020 ] 00:14:02.020 }' 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.020 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.281 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.281 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:02.281 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.281 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.281 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 0b5a2d29-feb2-4550-a784-854322ece7f7 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.540 [2024-10-29 11:03:07.847080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:02.540 NewBaseBdev 00:14:02.540 [2024-10-29 11:03:07.847392] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:02.540 [2024-10-29 11:03:07.847417] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:02.540 [2024-10-29 11:03:07.847692] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:14:02.540 [2024-10-29 11:03:07.848156] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:02.540 [2024-10-29 11:03:07.848169] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:14:02.540 [2024-10-29 11:03:07.848282] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:14:02.540 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.541 [ 00:14:02.541 { 00:14:02.541 "name": "NewBaseBdev", 00:14:02.541 "aliases": [ 00:14:02.541 "0b5a2d29-feb2-4550-a784-854322ece7f7" 00:14:02.541 ], 00:14:02.541 "product_name": "Malloc disk", 00:14:02.541 "block_size": 512, 00:14:02.541 "num_blocks": 65536, 00:14:02.541 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:02.541 "assigned_rate_limits": { 00:14:02.541 "rw_ios_per_sec": 0, 00:14:02.541 "rw_mbytes_per_sec": 0, 00:14:02.541 "r_mbytes_per_sec": 0, 00:14:02.541 "w_mbytes_per_sec": 0 00:14:02.541 }, 00:14:02.541 "claimed": true, 00:14:02.541 "claim_type": "exclusive_write", 00:14:02.541 "zoned": false, 00:14:02.541 "supported_io_types": { 00:14:02.541 "read": true, 00:14:02.541 "write": true, 00:14:02.541 "unmap": true, 00:14:02.541 "flush": true, 00:14:02.541 "reset": true, 00:14:02.541 "nvme_admin": false, 00:14:02.541 "nvme_io": false, 00:14:02.541 "nvme_io_md": false, 00:14:02.541 "write_zeroes": true, 00:14:02.541 "zcopy": true, 00:14:02.541 "get_zone_info": false, 00:14:02.541 "zone_management": false, 00:14:02.541 "zone_append": false, 00:14:02.541 "compare": false, 00:14:02.541 "compare_and_write": false, 00:14:02.541 "abort": true, 00:14:02.541 "seek_hole": false, 00:14:02.541 "seek_data": false, 00:14:02.541 "copy": true, 00:14:02.541 "nvme_iov_md": false 00:14:02.541 }, 00:14:02.541 "memory_domains": [ 00:14:02.541 { 00:14:02.541 "dma_device_id": "system", 00:14:02.541 "dma_device_type": 1 00:14:02.541 }, 00:14:02.541 { 00:14:02.541 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.541 "dma_device_type": 2 00:14:02.541 } 00:14:02.541 ], 00:14:02.541 "driver_specific": {} 00:14:02.541 } 00:14:02.541 ] 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.541 "name": "Existed_Raid", 00:14:02.541 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:02.541 "strip_size_kb": 64, 00:14:02.541 "state": "online", 00:14:02.541 "raid_level": "raid5f", 00:14:02.541 "superblock": true, 00:14:02.541 "num_base_bdevs": 3, 00:14:02.541 "num_base_bdevs_discovered": 3, 00:14:02.541 "num_base_bdevs_operational": 3, 00:14:02.541 "base_bdevs_list": [ 00:14:02.541 { 00:14:02.541 "name": "NewBaseBdev", 00:14:02.541 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:02.541 "is_configured": true, 00:14:02.541 "data_offset": 2048, 00:14:02.541 "data_size": 63488 00:14:02.541 }, 00:14:02.541 { 00:14:02.541 "name": "BaseBdev2", 00:14:02.541 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:02.541 "is_configured": true, 00:14:02.541 "data_offset": 2048, 00:14:02.541 "data_size": 63488 00:14:02.541 }, 00:14:02.541 { 00:14:02.541 "name": "BaseBdev3", 00:14:02.541 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:02.541 "is_configured": true, 00:14:02.541 "data_offset": 2048, 00:14:02.541 "data_size": 63488 00:14:02.541 } 00:14:02.541 ] 00:14:02.541 }' 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.541 11:03:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:03.109 [2024-10-29 11:03:08.338480] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:03.109 "name": "Existed_Raid", 00:14:03.109 "aliases": [ 00:14:03.109 "e37d1c93-7cec-4315-8194-5b200484056c" 00:14:03.109 ], 00:14:03.109 "product_name": "Raid Volume", 00:14:03.109 "block_size": 512, 00:14:03.109 "num_blocks": 126976, 00:14:03.109 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:03.109 "assigned_rate_limits": { 00:14:03.109 "rw_ios_per_sec": 0, 00:14:03.109 "rw_mbytes_per_sec": 0, 00:14:03.109 "r_mbytes_per_sec": 0, 00:14:03.109 "w_mbytes_per_sec": 0 00:14:03.109 }, 00:14:03.109 "claimed": false, 00:14:03.109 "zoned": false, 00:14:03.109 "supported_io_types": { 00:14:03.109 "read": true, 00:14:03.109 "write": true, 00:14:03.109 "unmap": false, 00:14:03.109 "flush": false, 00:14:03.109 "reset": true, 00:14:03.109 "nvme_admin": false, 00:14:03.109 "nvme_io": false, 00:14:03.109 "nvme_io_md": false, 00:14:03.109 "write_zeroes": true, 00:14:03.109 "zcopy": false, 00:14:03.109 "get_zone_info": false, 00:14:03.109 "zone_management": false, 00:14:03.109 "zone_append": false, 00:14:03.109 "compare": false, 00:14:03.109 "compare_and_write": false, 00:14:03.109 "abort": false, 00:14:03.109 "seek_hole": false, 00:14:03.109 "seek_data": false, 00:14:03.109 "copy": false, 00:14:03.109 "nvme_iov_md": false 00:14:03.109 }, 00:14:03.109 "driver_specific": { 00:14:03.109 "raid": { 00:14:03.109 "uuid": "e37d1c93-7cec-4315-8194-5b200484056c", 00:14:03.109 "strip_size_kb": 64, 00:14:03.109 "state": "online", 00:14:03.109 "raid_level": "raid5f", 00:14:03.109 "superblock": true, 00:14:03.109 "num_base_bdevs": 3, 00:14:03.109 "num_base_bdevs_discovered": 3, 00:14:03.109 "num_base_bdevs_operational": 3, 00:14:03.109 "base_bdevs_list": [ 00:14:03.109 { 00:14:03.109 "name": "NewBaseBdev", 00:14:03.109 "uuid": "0b5a2d29-feb2-4550-a784-854322ece7f7", 00:14:03.109 "is_configured": true, 00:14:03.109 "data_offset": 2048, 00:14:03.109 "data_size": 63488 00:14:03.109 }, 00:14:03.109 { 00:14:03.109 "name": "BaseBdev2", 00:14:03.109 "uuid": "3e8c8c4f-fdb0-4092-befd-ec3a38b2832c", 00:14:03.109 "is_configured": true, 00:14:03.109 "data_offset": 2048, 00:14:03.109 "data_size": 63488 00:14:03.109 }, 00:14:03.109 { 00:14:03.109 "name": "BaseBdev3", 00:14:03.109 "uuid": "ac0a43e2-e204-45de-b0dc-73a11a91de6a", 00:14:03.109 "is_configured": true, 00:14:03.109 "data_offset": 2048, 00:14:03.109 "data_size": 63488 00:14:03.109 } 00:14:03.109 ] 00:14:03.109 } 00:14:03.109 } 00:14:03.109 }' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:03.109 BaseBdev2 00:14:03.109 BaseBdev3' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.109 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.110 [2024-10-29 11:03:08.593834] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:03.110 [2024-10-29 11:03:08.593909] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:03.110 [2024-10-29 11:03:08.594019] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:03.110 [2024-10-29 11:03:08.594314] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:03.110 [2024-10-29 11:03:08.594416] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 91165 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 91165 ']' 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 91165 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:14:03.110 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 91165 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 91165' 00:14:03.369 killing process with pid 91165 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 91165 00:14:03.369 [2024-10-29 11:03:08.643285] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:03.369 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 91165 00:14:03.369 [2024-10-29 11:03:08.673589] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:03.630 11:03:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:03.630 00:14:03.630 real 0m8.435s 00:14:03.630 user 0m14.296s 00:14:03.630 sys 0m1.826s 00:14:03.630 ************************************ 00:14:03.630 END TEST raid5f_state_function_test_sb 00:14:03.630 ************************************ 00:14:03.630 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:14:03.630 11:03:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.630 11:03:08 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:14:03.630 11:03:08 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:14:03.630 11:03:08 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:14:03.630 11:03:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:03.630 ************************************ 00:14:03.630 START TEST raid5f_superblock_test 00:14:03.630 ************************************ 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid5f 3 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91769 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91769 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 91769 ']' 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:03.630 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:14:03.630 11:03:08 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:03.630 [2024-10-29 11:03:09.059595] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:14:03.630 [2024-10-29 11:03:09.059803] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91769 ] 00:14:03.890 [2024-10-29 11:03:09.231092] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:03.890 [2024-10-29 11:03:09.256347] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:03.890 [2024-10-29 11:03:09.299549] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:03.890 [2024-10-29 11:03:09.299589] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.459 malloc1 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.459 [2024-10-29 11:03:09.902805] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:04.459 [2024-10-29 11:03:09.902936] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:04.459 [2024-10-29 11:03:09.903002] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:04.459 [2024-10-29 11:03:09.903054] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:04.459 [2024-10-29 11:03:09.905226] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:04.459 [2024-10-29 11:03:09.905310] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:04.459 pt1 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.459 malloc2 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.459 [2024-10-29 11:03:09.935498] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:04.459 [2024-10-29 11:03:09.935557] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:04.459 [2024-10-29 11:03:09.935576] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:04.459 [2024-10-29 11:03:09.935604] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:04.459 [2024-10-29 11:03:09.937758] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:04.459 [2024-10-29 11:03:09.937802] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:04.459 pt2 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.459 malloc3 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.459 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.720 [2024-10-29 11:03:09.964416] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:04.720 [2024-10-29 11:03:09.964517] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:04.720 [2024-10-29 11:03:09.964569] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:04.720 [2024-10-29 11:03:09.964609] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:04.720 [2024-10-29 11:03:09.966799] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:04.720 [2024-10-29 11:03:09.966880] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:04.720 pt3 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.720 [2024-10-29 11:03:09.976418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:04.720 [2024-10-29 11:03:09.978293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:04.720 [2024-10-29 11:03:09.978410] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:04.720 [2024-10-29 11:03:09.978612] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:14:04.720 [2024-10-29 11:03:09.978665] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:04.720 [2024-10-29 11:03:09.978978] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:14:04.720 [2024-10-29 11:03:09.979485] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:14:04.720 [2024-10-29 11:03:09.979542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:14:04.720 [2024-10-29 11:03:09.979748] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.720 11:03:09 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.720 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.720 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.720 "name": "raid_bdev1", 00:14:04.720 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:04.720 "strip_size_kb": 64, 00:14:04.720 "state": "online", 00:14:04.720 "raid_level": "raid5f", 00:14:04.720 "superblock": true, 00:14:04.720 "num_base_bdevs": 3, 00:14:04.720 "num_base_bdevs_discovered": 3, 00:14:04.720 "num_base_bdevs_operational": 3, 00:14:04.720 "base_bdevs_list": [ 00:14:04.720 { 00:14:04.720 "name": "pt1", 00:14:04.720 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:04.720 "is_configured": true, 00:14:04.720 "data_offset": 2048, 00:14:04.720 "data_size": 63488 00:14:04.720 }, 00:14:04.720 { 00:14:04.720 "name": "pt2", 00:14:04.720 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:04.720 "is_configured": true, 00:14:04.720 "data_offset": 2048, 00:14:04.720 "data_size": 63488 00:14:04.720 }, 00:14:04.720 { 00:14:04.720 "name": "pt3", 00:14:04.720 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:04.720 "is_configured": true, 00:14:04.720 "data_offset": 2048, 00:14:04.720 "data_size": 63488 00:14:04.720 } 00:14:04.720 ] 00:14:04.720 }' 00:14:04.720 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.720 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:04.981 [2024-10-29 11:03:10.420638] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:04.981 "name": "raid_bdev1", 00:14:04.981 "aliases": [ 00:14:04.981 "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca" 00:14:04.981 ], 00:14:04.981 "product_name": "Raid Volume", 00:14:04.981 "block_size": 512, 00:14:04.981 "num_blocks": 126976, 00:14:04.981 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:04.981 "assigned_rate_limits": { 00:14:04.981 "rw_ios_per_sec": 0, 00:14:04.981 "rw_mbytes_per_sec": 0, 00:14:04.981 "r_mbytes_per_sec": 0, 00:14:04.981 "w_mbytes_per_sec": 0 00:14:04.981 }, 00:14:04.981 "claimed": false, 00:14:04.981 "zoned": false, 00:14:04.981 "supported_io_types": { 00:14:04.981 "read": true, 00:14:04.981 "write": true, 00:14:04.981 "unmap": false, 00:14:04.981 "flush": false, 00:14:04.981 "reset": true, 00:14:04.981 "nvme_admin": false, 00:14:04.981 "nvme_io": false, 00:14:04.981 "nvme_io_md": false, 00:14:04.981 "write_zeroes": true, 00:14:04.981 "zcopy": false, 00:14:04.981 "get_zone_info": false, 00:14:04.981 "zone_management": false, 00:14:04.981 "zone_append": false, 00:14:04.981 "compare": false, 00:14:04.981 "compare_and_write": false, 00:14:04.981 "abort": false, 00:14:04.981 "seek_hole": false, 00:14:04.981 "seek_data": false, 00:14:04.981 "copy": false, 00:14:04.981 "nvme_iov_md": false 00:14:04.981 }, 00:14:04.981 "driver_specific": { 00:14:04.981 "raid": { 00:14:04.981 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:04.981 "strip_size_kb": 64, 00:14:04.981 "state": "online", 00:14:04.981 "raid_level": "raid5f", 00:14:04.981 "superblock": true, 00:14:04.981 "num_base_bdevs": 3, 00:14:04.981 "num_base_bdevs_discovered": 3, 00:14:04.981 "num_base_bdevs_operational": 3, 00:14:04.981 "base_bdevs_list": [ 00:14:04.981 { 00:14:04.981 "name": "pt1", 00:14:04.981 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:04.981 "is_configured": true, 00:14:04.981 "data_offset": 2048, 00:14:04.981 "data_size": 63488 00:14:04.981 }, 00:14:04.981 { 00:14:04.981 "name": "pt2", 00:14:04.981 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:04.981 "is_configured": true, 00:14:04.981 "data_offset": 2048, 00:14:04.981 "data_size": 63488 00:14:04.981 }, 00:14:04.981 { 00:14:04.981 "name": "pt3", 00:14:04.981 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:04.981 "is_configured": true, 00:14:04.981 "data_offset": 2048, 00:14:04.981 "data_size": 63488 00:14:04.981 } 00:14:04.981 ] 00:14:04.981 } 00:14:04.981 } 00:14:04.981 }' 00:14:04.981 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:05.242 pt2 00:14:05.242 pt3' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.242 [2024-10-29 11:03:10.684129] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca ']' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.242 [2024-10-29 11:03:10.727910] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:05.242 [2024-10-29 11:03:10.727935] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:05.242 [2024-10-29 11:03:10.728004] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:05.242 [2024-10-29 11:03:10.728071] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:05.242 [2024-10-29 11:03:10.728093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:05.242 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 [2024-10-29 11:03:10.883716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:05.505 [2024-10-29 11:03:10.885550] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:05.505 [2024-10-29 11:03:10.885601] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:05.505 [2024-10-29 11:03:10.885651] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:05.505 [2024-10-29 11:03:10.885703] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:05.505 [2024-10-29 11:03:10.885729] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:05.505 [2024-10-29 11:03:10.885744] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:05.505 [2024-10-29 11:03:10.885758] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:14:05.505 request: 00:14:05.505 { 00:14:05.505 "name": "raid_bdev1", 00:14:05.505 "raid_level": "raid5f", 00:14:05.505 "base_bdevs": [ 00:14:05.505 "malloc1", 00:14:05.505 "malloc2", 00:14:05.505 "malloc3" 00:14:05.505 ], 00:14:05.505 "strip_size_kb": 64, 00:14:05.505 "superblock": false, 00:14:05.505 "method": "bdev_raid_create", 00:14:05.505 "req_id": 1 00:14:05.505 } 00:14:05.505 Got JSON-RPC error response 00:14:05.505 response: 00:14:05.505 { 00:14:05.505 "code": -17, 00:14:05.505 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:05.505 } 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.505 [2024-10-29 11:03:10.939625] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:05.505 [2024-10-29 11:03:10.939679] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:05.505 [2024-10-29 11:03:10.939697] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:05.505 [2024-10-29 11:03:10.939710] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:05.505 [2024-10-29 11:03:10.941785] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:05.505 [2024-10-29 11:03:10.941841] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:05.505 [2024-10-29 11:03:10.941910] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:05.505 [2024-10-29 11:03:10.941955] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:05.505 pt1 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:05.505 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.506 "name": "raid_bdev1", 00:14:05.506 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:05.506 "strip_size_kb": 64, 00:14:05.506 "state": "configuring", 00:14:05.506 "raid_level": "raid5f", 00:14:05.506 "superblock": true, 00:14:05.506 "num_base_bdevs": 3, 00:14:05.506 "num_base_bdevs_discovered": 1, 00:14:05.506 "num_base_bdevs_operational": 3, 00:14:05.506 "base_bdevs_list": [ 00:14:05.506 { 00:14:05.506 "name": "pt1", 00:14:05.506 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:05.506 "is_configured": true, 00:14:05.506 "data_offset": 2048, 00:14:05.506 "data_size": 63488 00:14:05.506 }, 00:14:05.506 { 00:14:05.506 "name": null, 00:14:05.506 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:05.506 "is_configured": false, 00:14:05.506 "data_offset": 2048, 00:14:05.506 "data_size": 63488 00:14:05.506 }, 00:14:05.506 { 00:14:05.506 "name": null, 00:14:05.506 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:05.506 "is_configured": false, 00:14:05.506 "data_offset": 2048, 00:14:05.506 "data_size": 63488 00:14:05.506 } 00:14:05.506 ] 00:14:05.506 }' 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.506 11:03:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.076 [2024-10-29 11:03:11.406806] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:06.076 [2024-10-29 11:03:11.406915] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.076 [2024-10-29 11:03:11.406954] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:06.076 [2024-10-29 11:03:11.406990] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.076 [2024-10-29 11:03:11.407484] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.076 [2024-10-29 11:03:11.407554] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:06.076 [2024-10-29 11:03:11.407680] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:06.076 [2024-10-29 11:03:11.407742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:06.076 pt2 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.076 [2024-10-29 11:03:11.418784] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.076 "name": "raid_bdev1", 00:14:06.076 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:06.076 "strip_size_kb": 64, 00:14:06.076 "state": "configuring", 00:14:06.076 "raid_level": "raid5f", 00:14:06.076 "superblock": true, 00:14:06.076 "num_base_bdevs": 3, 00:14:06.076 "num_base_bdevs_discovered": 1, 00:14:06.076 "num_base_bdevs_operational": 3, 00:14:06.076 "base_bdevs_list": [ 00:14:06.076 { 00:14:06.076 "name": "pt1", 00:14:06.076 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:06.076 "is_configured": true, 00:14:06.076 "data_offset": 2048, 00:14:06.076 "data_size": 63488 00:14:06.076 }, 00:14:06.076 { 00:14:06.076 "name": null, 00:14:06.076 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:06.076 "is_configured": false, 00:14:06.076 "data_offset": 0, 00:14:06.076 "data_size": 63488 00:14:06.076 }, 00:14:06.076 { 00:14:06.076 "name": null, 00:14:06.076 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:06.076 "is_configured": false, 00:14:06.076 "data_offset": 2048, 00:14:06.076 "data_size": 63488 00:14:06.076 } 00:14:06.076 ] 00:14:06.076 }' 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.076 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.336 [2024-10-29 11:03:11.798195] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:06.336 [2024-10-29 11:03:11.798254] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.336 [2024-10-29 11:03:11.798290] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:06.336 [2024-10-29 11:03:11.798300] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.336 [2024-10-29 11:03:11.798679] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.336 [2024-10-29 11:03:11.798698] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:06.336 [2024-10-29 11:03:11.798767] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:06.336 [2024-10-29 11:03:11.798787] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:06.336 pt2 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.336 [2024-10-29 11:03:11.806160] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:06.336 [2024-10-29 11:03:11.806210] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:06.336 [2024-10-29 11:03:11.806229] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:06.336 [2024-10-29 11:03:11.806238] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:06.336 [2024-10-29 11:03:11.806590] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:06.336 [2024-10-29 11:03:11.806609] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:06.336 [2024-10-29 11:03:11.806665] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:06.336 [2024-10-29 11:03:11.806684] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:06.336 [2024-10-29 11:03:11.806790] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:06.336 [2024-10-29 11:03:11.806801] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:06.336 [2024-10-29 11:03:11.807016] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:06.336 [2024-10-29 11:03:11.807468] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:06.336 [2024-10-29 11:03:11.807491] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:14:06.336 [2024-10-29 11:03:11.807596] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:06.336 pt3 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.336 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.595 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.595 "name": "raid_bdev1", 00:14:06.595 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:06.595 "strip_size_kb": 64, 00:14:06.595 "state": "online", 00:14:06.595 "raid_level": "raid5f", 00:14:06.595 "superblock": true, 00:14:06.595 "num_base_bdevs": 3, 00:14:06.595 "num_base_bdevs_discovered": 3, 00:14:06.595 "num_base_bdevs_operational": 3, 00:14:06.595 "base_bdevs_list": [ 00:14:06.595 { 00:14:06.595 "name": "pt1", 00:14:06.595 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:06.595 "is_configured": true, 00:14:06.595 "data_offset": 2048, 00:14:06.595 "data_size": 63488 00:14:06.595 }, 00:14:06.595 { 00:14:06.595 "name": "pt2", 00:14:06.595 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:06.595 "is_configured": true, 00:14:06.595 "data_offset": 2048, 00:14:06.595 "data_size": 63488 00:14:06.595 }, 00:14:06.595 { 00:14:06.595 "name": "pt3", 00:14:06.595 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:06.595 "is_configured": true, 00:14:06.595 "data_offset": 2048, 00:14:06.595 "data_size": 63488 00:14:06.595 } 00:14:06.595 ] 00:14:06.595 }' 00:14:06.595 11:03:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.595 11:03:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:06.855 [2024-10-29 11:03:12.209698] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.855 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:06.855 "name": "raid_bdev1", 00:14:06.855 "aliases": [ 00:14:06.855 "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca" 00:14:06.855 ], 00:14:06.855 "product_name": "Raid Volume", 00:14:06.855 "block_size": 512, 00:14:06.855 "num_blocks": 126976, 00:14:06.855 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:06.855 "assigned_rate_limits": { 00:14:06.855 "rw_ios_per_sec": 0, 00:14:06.855 "rw_mbytes_per_sec": 0, 00:14:06.855 "r_mbytes_per_sec": 0, 00:14:06.855 "w_mbytes_per_sec": 0 00:14:06.855 }, 00:14:06.855 "claimed": false, 00:14:06.855 "zoned": false, 00:14:06.855 "supported_io_types": { 00:14:06.855 "read": true, 00:14:06.855 "write": true, 00:14:06.855 "unmap": false, 00:14:06.855 "flush": false, 00:14:06.855 "reset": true, 00:14:06.855 "nvme_admin": false, 00:14:06.855 "nvme_io": false, 00:14:06.855 "nvme_io_md": false, 00:14:06.855 "write_zeroes": true, 00:14:06.855 "zcopy": false, 00:14:06.855 "get_zone_info": false, 00:14:06.855 "zone_management": false, 00:14:06.855 "zone_append": false, 00:14:06.855 "compare": false, 00:14:06.855 "compare_and_write": false, 00:14:06.855 "abort": false, 00:14:06.855 "seek_hole": false, 00:14:06.855 "seek_data": false, 00:14:06.855 "copy": false, 00:14:06.855 "nvme_iov_md": false 00:14:06.855 }, 00:14:06.855 "driver_specific": { 00:14:06.855 "raid": { 00:14:06.855 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:06.855 "strip_size_kb": 64, 00:14:06.855 "state": "online", 00:14:06.855 "raid_level": "raid5f", 00:14:06.855 "superblock": true, 00:14:06.855 "num_base_bdevs": 3, 00:14:06.855 "num_base_bdevs_discovered": 3, 00:14:06.856 "num_base_bdevs_operational": 3, 00:14:06.856 "base_bdevs_list": [ 00:14:06.856 { 00:14:06.856 "name": "pt1", 00:14:06.856 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:06.856 "is_configured": true, 00:14:06.856 "data_offset": 2048, 00:14:06.856 "data_size": 63488 00:14:06.856 }, 00:14:06.856 { 00:14:06.856 "name": "pt2", 00:14:06.856 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:06.856 "is_configured": true, 00:14:06.856 "data_offset": 2048, 00:14:06.856 "data_size": 63488 00:14:06.856 }, 00:14:06.856 { 00:14:06.856 "name": "pt3", 00:14:06.856 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:06.856 "is_configured": true, 00:14:06.856 "data_offset": 2048, 00:14:06.856 "data_size": 63488 00:14:06.856 } 00:14:06.856 ] 00:14:06.856 } 00:14:06.856 } 00:14:06.856 }' 00:14:06.856 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:06.856 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:06.856 pt2 00:14:06.856 pt3' 00:14:06.856 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:06.856 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:06.856 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 [2024-10-29 11:03:12.513168] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca '!=' a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca ']' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 [2024-10-29 11:03:12.540994] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.116 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.116 "name": "raid_bdev1", 00:14:07.116 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:07.116 "strip_size_kb": 64, 00:14:07.116 "state": "online", 00:14:07.116 "raid_level": "raid5f", 00:14:07.116 "superblock": true, 00:14:07.116 "num_base_bdevs": 3, 00:14:07.116 "num_base_bdevs_discovered": 2, 00:14:07.116 "num_base_bdevs_operational": 2, 00:14:07.116 "base_bdevs_list": [ 00:14:07.116 { 00:14:07.116 "name": null, 00:14:07.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.117 "is_configured": false, 00:14:07.117 "data_offset": 0, 00:14:07.117 "data_size": 63488 00:14:07.117 }, 00:14:07.117 { 00:14:07.117 "name": "pt2", 00:14:07.117 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:07.117 "is_configured": true, 00:14:07.117 "data_offset": 2048, 00:14:07.117 "data_size": 63488 00:14:07.117 }, 00:14:07.117 { 00:14:07.117 "name": "pt3", 00:14:07.117 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:07.117 "is_configured": true, 00:14:07.117 "data_offset": 2048, 00:14:07.117 "data_size": 63488 00:14:07.117 } 00:14:07.117 ] 00:14:07.117 }' 00:14:07.117 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.117 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.376 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:07.376 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.376 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.637 [2024-10-29 11:03:12.880461] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:07.637 [2024-10-29 11:03:12.880538] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:07.637 [2024-10-29 11:03:12.880618] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:07.637 [2024-10-29 11:03:12.880720] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:07.637 [2024-10-29 11:03:12.880780] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:07.637 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.638 [2024-10-29 11:03:12.968314] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:07.638 [2024-10-29 11:03:12.968425] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:07.638 [2024-10-29 11:03:12.968468] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:14:07.638 [2024-10-29 11:03:12.968479] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:07.638 [2024-10-29 11:03:12.970693] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:07.638 [2024-10-29 11:03:12.970733] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:07.638 [2024-10-29 11:03:12.970804] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:07.638 [2024-10-29 11:03:12.970838] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:07.638 pt2 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.638 11:03:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.638 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:07.638 "name": "raid_bdev1", 00:14:07.638 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:07.638 "strip_size_kb": 64, 00:14:07.638 "state": "configuring", 00:14:07.638 "raid_level": "raid5f", 00:14:07.638 "superblock": true, 00:14:07.638 "num_base_bdevs": 3, 00:14:07.638 "num_base_bdevs_discovered": 1, 00:14:07.638 "num_base_bdevs_operational": 2, 00:14:07.638 "base_bdevs_list": [ 00:14:07.638 { 00:14:07.638 "name": null, 00:14:07.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:07.638 "is_configured": false, 00:14:07.638 "data_offset": 2048, 00:14:07.638 "data_size": 63488 00:14:07.638 }, 00:14:07.638 { 00:14:07.638 "name": "pt2", 00:14:07.638 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:07.638 "is_configured": true, 00:14:07.638 "data_offset": 2048, 00:14:07.638 "data_size": 63488 00:14:07.638 }, 00:14:07.638 { 00:14:07.638 "name": null, 00:14:07.638 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:07.638 "is_configured": false, 00:14:07.638 "data_offset": 2048, 00:14:07.638 "data_size": 63488 00:14:07.638 } 00:14:07.638 ] 00:14:07.638 }' 00:14:07.638 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:07.638 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.207 [2024-10-29 11:03:13.427588] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:08.207 [2024-10-29 11:03:13.427698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.207 [2024-10-29 11:03:13.427742] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:08.207 [2024-10-29 11:03:13.427776] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.207 [2024-10-29 11:03:13.428275] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.207 [2024-10-29 11:03:13.428347] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:08.207 [2024-10-29 11:03:13.428506] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:08.207 [2024-10-29 11:03:13.428593] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:08.207 [2024-10-29 11:03:13.428752] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:08.207 [2024-10-29 11:03:13.428796] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:08.207 [2024-10-29 11:03:13.429093] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:14:08.207 [2024-10-29 11:03:13.429656] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:08.207 [2024-10-29 11:03:13.429716] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:14:08.207 [2024-10-29 11:03:13.430015] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:08.207 pt3 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.207 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:08.207 "name": "raid_bdev1", 00:14:08.207 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:08.207 "strip_size_kb": 64, 00:14:08.207 "state": "online", 00:14:08.207 "raid_level": "raid5f", 00:14:08.207 "superblock": true, 00:14:08.207 "num_base_bdevs": 3, 00:14:08.207 "num_base_bdevs_discovered": 2, 00:14:08.207 "num_base_bdevs_operational": 2, 00:14:08.207 "base_bdevs_list": [ 00:14:08.207 { 00:14:08.207 "name": null, 00:14:08.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.207 "is_configured": false, 00:14:08.207 "data_offset": 2048, 00:14:08.207 "data_size": 63488 00:14:08.207 }, 00:14:08.207 { 00:14:08.207 "name": "pt2", 00:14:08.207 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:08.207 "is_configured": true, 00:14:08.207 "data_offset": 2048, 00:14:08.207 "data_size": 63488 00:14:08.207 }, 00:14:08.207 { 00:14:08.207 "name": "pt3", 00:14:08.207 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:08.207 "is_configured": true, 00:14:08.207 "data_offset": 2048, 00:14:08.207 "data_size": 63488 00:14:08.207 } 00:14:08.208 ] 00:14:08.208 }' 00:14:08.208 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:08.208 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 [2024-10-29 11:03:13.826920] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:08.468 [2024-10-29 11:03:13.826947] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:08.468 [2024-10-29 11:03:13.827014] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:08.468 [2024-10-29 11:03:13.827082] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:08.468 [2024-10-29 11:03:13.827096] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 [2024-10-29 11:03:13.902779] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:08.468 [2024-10-29 11:03:13.902838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.468 [2024-10-29 11:03:13.902854] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:08.468 [2024-10-29 11:03:13.902866] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.468 [2024-10-29 11:03:13.905065] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.468 [2024-10-29 11:03:13.905110] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:08.468 [2024-10-29 11:03:13.905174] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:08.468 [2024-10-29 11:03:13.905218] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:08.468 [2024-10-29 11:03:13.905331] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:08.468 [2024-10-29 11:03:13.905355] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:08.468 [2024-10-29 11:03:13.905397] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:14:08.468 [2024-10-29 11:03:13.905446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:08.468 pt1 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.468 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:08.468 "name": "raid_bdev1", 00:14:08.468 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:08.468 "strip_size_kb": 64, 00:14:08.468 "state": "configuring", 00:14:08.468 "raid_level": "raid5f", 00:14:08.468 "superblock": true, 00:14:08.468 "num_base_bdevs": 3, 00:14:08.468 "num_base_bdevs_discovered": 1, 00:14:08.468 "num_base_bdevs_operational": 2, 00:14:08.468 "base_bdevs_list": [ 00:14:08.468 { 00:14:08.468 "name": null, 00:14:08.468 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.468 "is_configured": false, 00:14:08.468 "data_offset": 2048, 00:14:08.468 "data_size": 63488 00:14:08.468 }, 00:14:08.468 { 00:14:08.468 "name": "pt2", 00:14:08.468 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:08.468 "is_configured": true, 00:14:08.468 "data_offset": 2048, 00:14:08.468 "data_size": 63488 00:14:08.468 }, 00:14:08.468 { 00:14:08.468 "name": null, 00:14:08.468 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:08.469 "is_configured": false, 00:14:08.469 "data_offset": 2048, 00:14:08.469 "data_size": 63488 00:14:08.469 } 00:14:08.469 ] 00:14:08.469 }' 00:14:08.469 11:03:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:08.469 11:03:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:09.038 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.039 [2024-10-29 11:03:14.398084] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:09.039 [2024-10-29 11:03:14.398212] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:09.039 [2024-10-29 11:03:14.398274] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:09.039 [2024-10-29 11:03:14.398320] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:09.039 [2024-10-29 11:03:14.398783] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:09.039 [2024-10-29 11:03:14.398858] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:09.039 [2024-10-29 11:03:14.398980] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:09.039 [2024-10-29 11:03:14.399043] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:09.039 [2024-10-29 11:03:14.399191] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:14:09.039 [2024-10-29 11:03:14.399239] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:09.039 [2024-10-29 11:03:14.399544] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:14:09.039 [2024-10-29 11:03:14.400094] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:14:09.039 [2024-10-29 11:03:14.400154] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:14:09.039 [2024-10-29 11:03:14.400431] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:09.039 pt3 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.039 "name": "raid_bdev1", 00:14:09.039 "uuid": "a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca", 00:14:09.039 "strip_size_kb": 64, 00:14:09.039 "state": "online", 00:14:09.039 "raid_level": "raid5f", 00:14:09.039 "superblock": true, 00:14:09.039 "num_base_bdevs": 3, 00:14:09.039 "num_base_bdevs_discovered": 2, 00:14:09.039 "num_base_bdevs_operational": 2, 00:14:09.039 "base_bdevs_list": [ 00:14:09.039 { 00:14:09.039 "name": null, 00:14:09.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.039 "is_configured": false, 00:14:09.039 "data_offset": 2048, 00:14:09.039 "data_size": 63488 00:14:09.039 }, 00:14:09.039 { 00:14:09.039 "name": "pt2", 00:14:09.039 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:09.039 "is_configured": true, 00:14:09.039 "data_offset": 2048, 00:14:09.039 "data_size": 63488 00:14:09.039 }, 00:14:09.039 { 00:14:09.039 "name": "pt3", 00:14:09.039 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:09.039 "is_configured": true, 00:14:09.039 "data_offset": 2048, 00:14:09.039 "data_size": 63488 00:14:09.039 } 00:14:09.039 ] 00:14:09.039 }' 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.039 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.609 [2024-10-29 11:03:14.897458] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca '!=' a075fd7f-cf4f-40bd-b8f1-a2655a98b7ca ']' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91769 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 91769 ']' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # kill -0 91769 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@957 -- # uname 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 91769 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 91769' 00:14:09.609 killing process with pid 91769 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@971 -- # kill 91769 00:14:09.609 [2024-10-29 11:03:14.977910] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:09.609 [2024-10-29 11:03:14.977995] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:09.609 [2024-10-29 11:03:14.978061] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:09.609 [2024-10-29 11:03:14.978071] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:14:09.609 11:03:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@976 -- # wait 91769 00:14:09.609 [2024-10-29 11:03:15.010713] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:09.870 11:03:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:09.870 00:14:09.870 real 0m6.260s 00:14:09.870 user 0m10.412s 00:14:09.870 sys 0m1.352s 00:14:09.870 11:03:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:14:09.870 ************************************ 00:14:09.870 END TEST raid5f_superblock_test 00:14:09.870 ************************************ 00:14:09.870 11:03:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.870 11:03:15 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:09.870 11:03:15 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:14:09.870 11:03:15 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:14:09.870 11:03:15 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:14:09.870 11:03:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:09.870 ************************************ 00:14:09.870 START TEST raid5f_rebuild_test 00:14:09.870 ************************************ 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid5f 3 false false true 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=92191 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 92191 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@833 -- # '[' -z 92191 ']' 00:14:09.870 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:14:09.870 11:03:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.130 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:10.130 Zero copy mechanism will not be used. 00:14:10.130 [2024-10-29 11:03:15.419738] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:14:10.130 [2024-10-29 11:03:15.419879] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92191 ] 00:14:10.130 [2024-10-29 11:03:15.601399] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:10.130 [2024-10-29 11:03:15.628302] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:10.390 [2024-10-29 11:03:15.672485] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:10.390 [2024-10-29 11:03:15.672622] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@866 -- # return 0 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 BaseBdev1_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 [2024-10-29 11:03:16.260221] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:11.062 [2024-10-29 11:03:16.260302] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:11.062 [2024-10-29 11:03:16.260332] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:11.062 [2024-10-29 11:03:16.260349] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:11.062 [2024-10-29 11:03:16.262502] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:11.062 [2024-10-29 11:03:16.262591] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:11.062 BaseBdev1 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 BaseBdev2_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 [2024-10-29 11:03:16.288995] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:11.062 [2024-10-29 11:03:16.289062] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:11.062 [2024-10-29 11:03:16.289102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:11.062 [2024-10-29 11:03:16.289112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:11.062 [2024-10-29 11:03:16.291193] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:11.062 [2024-10-29 11:03:16.291232] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:11.062 BaseBdev2 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 BaseBdev3_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 [2024-10-29 11:03:16.317863] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:11.062 [2024-10-29 11:03:16.317918] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:11.062 [2024-10-29 11:03:16.317962] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:11.062 [2024-10-29 11:03:16.317972] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:11.062 [2024-10-29 11:03:16.320009] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:11.062 [2024-10-29 11:03:16.320049] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:11.062 BaseBdev3 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 spare_malloc 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 spare_delay 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.062 [2024-10-29 11:03:16.367816] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:11.062 [2024-10-29 11:03:16.367912] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:11.062 [2024-10-29 11:03:16.367958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:11.062 [2024-10-29 11:03:16.367969] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:11.062 [2024-10-29 11:03:16.370074] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:11.062 [2024-10-29 11:03:16.370112] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:11.062 spare 00:14:11.062 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.063 [2024-10-29 11:03:16.379865] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:11.063 [2024-10-29 11:03:16.381679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:11.063 [2024-10-29 11:03:16.381744] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:11.063 [2024-10-29 11:03:16.381826] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:14:11.063 [2024-10-29 11:03:16.381837] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:11.063 [2024-10-29 11:03:16.382097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:11.063 [2024-10-29 11:03:16.382524] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:14:11.063 [2024-10-29 11:03:16.382542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:14:11.063 [2024-10-29 11:03:16.382672] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.063 "name": "raid_bdev1", 00:14:11.063 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:11.063 "strip_size_kb": 64, 00:14:11.063 "state": "online", 00:14:11.063 "raid_level": "raid5f", 00:14:11.063 "superblock": false, 00:14:11.063 "num_base_bdevs": 3, 00:14:11.063 "num_base_bdevs_discovered": 3, 00:14:11.063 "num_base_bdevs_operational": 3, 00:14:11.063 "base_bdevs_list": [ 00:14:11.063 { 00:14:11.063 "name": "BaseBdev1", 00:14:11.063 "uuid": "7bd7b007-0a02-55cc-8541-7ce290975ede", 00:14:11.063 "is_configured": true, 00:14:11.063 "data_offset": 0, 00:14:11.063 "data_size": 65536 00:14:11.063 }, 00:14:11.063 { 00:14:11.063 "name": "BaseBdev2", 00:14:11.063 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:11.063 "is_configured": true, 00:14:11.063 "data_offset": 0, 00:14:11.063 "data_size": 65536 00:14:11.063 }, 00:14:11.063 { 00:14:11.063 "name": "BaseBdev3", 00:14:11.063 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:11.063 "is_configured": true, 00:14:11.063 "data_offset": 0, 00:14:11.063 "data_size": 65536 00:14:11.063 } 00:14:11.063 ] 00:14:11.063 }' 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.063 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.323 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:11.323 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.323 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.323 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:11.323 [2024-10-29 11:03:16.799779] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:11.323 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.582 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:14:11.582 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.582 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:11.582 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.582 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:11.583 11:03:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:11.583 [2024-10-29 11:03:17.047265] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:14:11.583 /dev/nbd0 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:11.842 1+0 records in 00:14:11.842 1+0 records out 00:14:11.842 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000400079 s, 10.2 MB/s 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:11.842 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:14:12.102 512+0 records in 00:14:12.102 512+0 records out 00:14:12.102 67108864 bytes (67 MB, 64 MiB) copied, 0.311108 s, 216 MB/s 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:12.102 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:12.362 [2024-10-29 11:03:17.652591] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.362 [2024-10-29 11:03:17.664851] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.362 "name": "raid_bdev1", 00:14:12.362 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:12.362 "strip_size_kb": 64, 00:14:12.362 "state": "online", 00:14:12.362 "raid_level": "raid5f", 00:14:12.362 "superblock": false, 00:14:12.362 "num_base_bdevs": 3, 00:14:12.362 "num_base_bdevs_discovered": 2, 00:14:12.362 "num_base_bdevs_operational": 2, 00:14:12.362 "base_bdevs_list": [ 00:14:12.362 { 00:14:12.362 "name": null, 00:14:12.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.362 "is_configured": false, 00:14:12.362 "data_offset": 0, 00:14:12.362 "data_size": 65536 00:14:12.362 }, 00:14:12.362 { 00:14:12.362 "name": "BaseBdev2", 00:14:12.362 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:12.362 "is_configured": true, 00:14:12.362 "data_offset": 0, 00:14:12.362 "data_size": 65536 00:14:12.362 }, 00:14:12.362 { 00:14:12.362 "name": "BaseBdev3", 00:14:12.362 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:12.362 "is_configured": true, 00:14:12.362 "data_offset": 0, 00:14:12.362 "data_size": 65536 00:14:12.362 } 00:14:12.362 ] 00:14:12.362 }' 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.362 11:03:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.931 11:03:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:12.931 11:03:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.932 11:03:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.932 [2024-10-29 11:03:18.139969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:12.932 [2024-10-29 11:03:18.144913] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b4e0 00:14:12.932 11:03:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.932 11:03:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:12.932 [2024-10-29 11:03:18.147152] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:13.871 "name": "raid_bdev1", 00:14:13.871 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:13.871 "strip_size_kb": 64, 00:14:13.871 "state": "online", 00:14:13.871 "raid_level": "raid5f", 00:14:13.871 "superblock": false, 00:14:13.871 "num_base_bdevs": 3, 00:14:13.871 "num_base_bdevs_discovered": 3, 00:14:13.871 "num_base_bdevs_operational": 3, 00:14:13.871 "process": { 00:14:13.871 "type": "rebuild", 00:14:13.871 "target": "spare", 00:14:13.871 "progress": { 00:14:13.871 "blocks": 20480, 00:14:13.871 "percent": 15 00:14:13.871 } 00:14:13.871 }, 00:14:13.871 "base_bdevs_list": [ 00:14:13.871 { 00:14:13.871 "name": "spare", 00:14:13.871 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:13.871 "is_configured": true, 00:14:13.871 "data_offset": 0, 00:14:13.871 "data_size": 65536 00:14:13.871 }, 00:14:13.871 { 00:14:13.871 "name": "BaseBdev2", 00:14:13.871 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:13.871 "is_configured": true, 00:14:13.871 "data_offset": 0, 00:14:13.871 "data_size": 65536 00:14:13.871 }, 00:14:13.871 { 00:14:13.871 "name": "BaseBdev3", 00:14:13.871 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:13.871 "is_configured": true, 00:14:13.871 "data_offset": 0, 00:14:13.871 "data_size": 65536 00:14:13.871 } 00:14:13.871 ] 00:14:13.871 }' 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.871 [2024-10-29 11:03:19.267011] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:13.871 [2024-10-29 11:03:19.354665] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:13.871 [2024-10-29 11:03:19.354747] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.871 [2024-10-29 11:03:19.354766] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:13.871 [2024-10-29 11:03:19.354778] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.871 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.872 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.872 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.131 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.131 "name": "raid_bdev1", 00:14:14.131 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:14.131 "strip_size_kb": 64, 00:14:14.131 "state": "online", 00:14:14.131 "raid_level": "raid5f", 00:14:14.131 "superblock": false, 00:14:14.131 "num_base_bdevs": 3, 00:14:14.131 "num_base_bdevs_discovered": 2, 00:14:14.131 "num_base_bdevs_operational": 2, 00:14:14.131 "base_bdevs_list": [ 00:14:14.131 { 00:14:14.131 "name": null, 00:14:14.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.131 "is_configured": false, 00:14:14.131 "data_offset": 0, 00:14:14.131 "data_size": 65536 00:14:14.131 }, 00:14:14.131 { 00:14:14.131 "name": "BaseBdev2", 00:14:14.131 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:14.131 "is_configured": true, 00:14:14.131 "data_offset": 0, 00:14:14.131 "data_size": 65536 00:14:14.131 }, 00:14:14.131 { 00:14:14.131 "name": "BaseBdev3", 00:14:14.131 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:14.131 "is_configured": true, 00:14:14.131 "data_offset": 0, 00:14:14.131 "data_size": 65536 00:14:14.131 } 00:14:14.131 ] 00:14:14.131 }' 00:14:14.132 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.132 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.393 "name": "raid_bdev1", 00:14:14.393 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:14.393 "strip_size_kb": 64, 00:14:14.393 "state": "online", 00:14:14.393 "raid_level": "raid5f", 00:14:14.393 "superblock": false, 00:14:14.393 "num_base_bdevs": 3, 00:14:14.393 "num_base_bdevs_discovered": 2, 00:14:14.393 "num_base_bdevs_operational": 2, 00:14:14.393 "base_bdevs_list": [ 00:14:14.393 { 00:14:14.393 "name": null, 00:14:14.393 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.393 "is_configured": false, 00:14:14.393 "data_offset": 0, 00:14:14.393 "data_size": 65536 00:14:14.393 }, 00:14:14.393 { 00:14:14.393 "name": "BaseBdev2", 00:14:14.393 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:14.393 "is_configured": true, 00:14:14.393 "data_offset": 0, 00:14:14.393 "data_size": 65536 00:14:14.393 }, 00:14:14.393 { 00:14:14.393 "name": "BaseBdev3", 00:14:14.393 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:14.393 "is_configured": true, 00:14:14.393 "data_offset": 0, 00:14:14.393 "data_size": 65536 00:14:14.393 } 00:14:14.393 ] 00:14:14.393 }' 00:14:14.393 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.653 [2024-10-29 11:03:19.979868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:14.653 [2024-10-29 11:03:19.984575] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b5b0 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.653 11:03:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:14.653 [2024-10-29 11:03:19.986768] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.592 11:03:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.592 11:03:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.592 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:15.592 "name": "raid_bdev1", 00:14:15.592 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:15.592 "strip_size_kb": 64, 00:14:15.592 "state": "online", 00:14:15.592 "raid_level": "raid5f", 00:14:15.592 "superblock": false, 00:14:15.592 "num_base_bdevs": 3, 00:14:15.592 "num_base_bdevs_discovered": 3, 00:14:15.592 "num_base_bdevs_operational": 3, 00:14:15.592 "process": { 00:14:15.592 "type": "rebuild", 00:14:15.592 "target": "spare", 00:14:15.592 "progress": { 00:14:15.592 "blocks": 20480, 00:14:15.592 "percent": 15 00:14:15.592 } 00:14:15.592 }, 00:14:15.592 "base_bdevs_list": [ 00:14:15.592 { 00:14:15.592 "name": "spare", 00:14:15.592 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:15.592 "is_configured": true, 00:14:15.592 "data_offset": 0, 00:14:15.592 "data_size": 65536 00:14:15.592 }, 00:14:15.592 { 00:14:15.592 "name": "BaseBdev2", 00:14:15.592 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:15.592 "is_configured": true, 00:14:15.592 "data_offset": 0, 00:14:15.592 "data_size": 65536 00:14:15.592 }, 00:14:15.592 { 00:14:15.592 "name": "BaseBdev3", 00:14:15.592 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:15.592 "is_configured": true, 00:14:15.592 "data_offset": 0, 00:14:15.592 "data_size": 65536 00:14:15.592 } 00:14:15.592 ] 00:14:15.593 }' 00:14:15.593 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=446 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.853 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:15.853 "name": "raid_bdev1", 00:14:15.853 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:15.853 "strip_size_kb": 64, 00:14:15.853 "state": "online", 00:14:15.853 "raid_level": "raid5f", 00:14:15.853 "superblock": false, 00:14:15.853 "num_base_bdevs": 3, 00:14:15.853 "num_base_bdevs_discovered": 3, 00:14:15.854 "num_base_bdevs_operational": 3, 00:14:15.854 "process": { 00:14:15.854 "type": "rebuild", 00:14:15.854 "target": "spare", 00:14:15.854 "progress": { 00:14:15.854 "blocks": 22528, 00:14:15.854 "percent": 17 00:14:15.854 } 00:14:15.854 }, 00:14:15.854 "base_bdevs_list": [ 00:14:15.854 { 00:14:15.854 "name": "spare", 00:14:15.854 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:15.854 "is_configured": true, 00:14:15.854 "data_offset": 0, 00:14:15.854 "data_size": 65536 00:14:15.854 }, 00:14:15.854 { 00:14:15.854 "name": "BaseBdev2", 00:14:15.854 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:15.854 "is_configured": true, 00:14:15.854 "data_offset": 0, 00:14:15.854 "data_size": 65536 00:14:15.854 }, 00:14:15.854 { 00:14:15.854 "name": "BaseBdev3", 00:14:15.854 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:15.854 "is_configured": true, 00:14:15.854 "data_offset": 0, 00:14:15.854 "data_size": 65536 00:14:15.854 } 00:14:15.854 ] 00:14:15.854 }' 00:14:15.854 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:15.854 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:15.854 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:15.854 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:15.854 11:03:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:16.794 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:16.794 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.795 11:03:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.055 11:03:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.055 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:17.055 "name": "raid_bdev1", 00:14:17.055 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:17.055 "strip_size_kb": 64, 00:14:17.055 "state": "online", 00:14:17.055 "raid_level": "raid5f", 00:14:17.055 "superblock": false, 00:14:17.055 "num_base_bdevs": 3, 00:14:17.055 "num_base_bdevs_discovered": 3, 00:14:17.055 "num_base_bdevs_operational": 3, 00:14:17.055 "process": { 00:14:17.055 "type": "rebuild", 00:14:17.055 "target": "spare", 00:14:17.055 "progress": { 00:14:17.055 "blocks": 45056, 00:14:17.055 "percent": 34 00:14:17.055 } 00:14:17.055 }, 00:14:17.055 "base_bdevs_list": [ 00:14:17.055 { 00:14:17.055 "name": "spare", 00:14:17.056 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:17.056 "is_configured": true, 00:14:17.056 "data_offset": 0, 00:14:17.056 "data_size": 65536 00:14:17.056 }, 00:14:17.056 { 00:14:17.056 "name": "BaseBdev2", 00:14:17.056 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:17.056 "is_configured": true, 00:14:17.056 "data_offset": 0, 00:14:17.056 "data_size": 65536 00:14:17.056 }, 00:14:17.056 { 00:14:17.056 "name": "BaseBdev3", 00:14:17.056 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:17.056 "is_configured": true, 00:14:17.056 "data_offset": 0, 00:14:17.056 "data_size": 65536 00:14:17.056 } 00:14:17.056 ] 00:14:17.056 }' 00:14:17.056 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:17.056 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:17.056 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:17.056 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:17.056 11:03:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:17.996 "name": "raid_bdev1", 00:14:17.996 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:17.996 "strip_size_kb": 64, 00:14:17.996 "state": "online", 00:14:17.996 "raid_level": "raid5f", 00:14:17.996 "superblock": false, 00:14:17.996 "num_base_bdevs": 3, 00:14:17.996 "num_base_bdevs_discovered": 3, 00:14:17.996 "num_base_bdevs_operational": 3, 00:14:17.996 "process": { 00:14:17.996 "type": "rebuild", 00:14:17.996 "target": "spare", 00:14:17.996 "progress": { 00:14:17.996 "blocks": 69632, 00:14:17.996 "percent": 53 00:14:17.996 } 00:14:17.996 }, 00:14:17.996 "base_bdevs_list": [ 00:14:17.996 { 00:14:17.996 "name": "spare", 00:14:17.996 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:17.996 "is_configured": true, 00:14:17.996 "data_offset": 0, 00:14:17.996 "data_size": 65536 00:14:17.996 }, 00:14:17.996 { 00:14:17.996 "name": "BaseBdev2", 00:14:17.996 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:17.996 "is_configured": true, 00:14:17.996 "data_offset": 0, 00:14:17.996 "data_size": 65536 00:14:17.996 }, 00:14:17.996 { 00:14:17.996 "name": "BaseBdev3", 00:14:17.996 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:17.996 "is_configured": true, 00:14:17.996 "data_offset": 0, 00:14:17.996 "data_size": 65536 00:14:17.996 } 00:14:17.996 ] 00:14:17.996 }' 00:14:17.996 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.256 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:18.256 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.256 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:18.256 11:03:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.197 "name": "raid_bdev1", 00:14:19.197 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:19.197 "strip_size_kb": 64, 00:14:19.197 "state": "online", 00:14:19.197 "raid_level": "raid5f", 00:14:19.197 "superblock": false, 00:14:19.197 "num_base_bdevs": 3, 00:14:19.197 "num_base_bdevs_discovered": 3, 00:14:19.197 "num_base_bdevs_operational": 3, 00:14:19.197 "process": { 00:14:19.197 "type": "rebuild", 00:14:19.197 "target": "spare", 00:14:19.197 "progress": { 00:14:19.197 "blocks": 92160, 00:14:19.197 "percent": 70 00:14:19.197 } 00:14:19.197 }, 00:14:19.197 "base_bdevs_list": [ 00:14:19.197 { 00:14:19.197 "name": "spare", 00:14:19.197 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:19.197 "is_configured": true, 00:14:19.197 "data_offset": 0, 00:14:19.197 "data_size": 65536 00:14:19.197 }, 00:14:19.197 { 00:14:19.197 "name": "BaseBdev2", 00:14:19.197 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:19.197 "is_configured": true, 00:14:19.197 "data_offset": 0, 00:14:19.197 "data_size": 65536 00:14:19.197 }, 00:14:19.197 { 00:14:19.197 "name": "BaseBdev3", 00:14:19.197 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:19.197 "is_configured": true, 00:14:19.197 "data_offset": 0, 00:14:19.197 "data_size": 65536 00:14:19.197 } 00:14:19.197 ] 00:14:19.197 }' 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.197 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.457 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.457 11:03:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.398 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:20.398 "name": "raid_bdev1", 00:14:20.398 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:20.398 "strip_size_kb": 64, 00:14:20.398 "state": "online", 00:14:20.398 "raid_level": "raid5f", 00:14:20.398 "superblock": false, 00:14:20.398 "num_base_bdevs": 3, 00:14:20.398 "num_base_bdevs_discovered": 3, 00:14:20.398 "num_base_bdevs_operational": 3, 00:14:20.398 "process": { 00:14:20.398 "type": "rebuild", 00:14:20.398 "target": "spare", 00:14:20.398 "progress": { 00:14:20.398 "blocks": 116736, 00:14:20.398 "percent": 89 00:14:20.398 } 00:14:20.398 }, 00:14:20.398 "base_bdevs_list": [ 00:14:20.398 { 00:14:20.398 "name": "spare", 00:14:20.398 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:20.398 "is_configured": true, 00:14:20.398 "data_offset": 0, 00:14:20.398 "data_size": 65536 00:14:20.398 }, 00:14:20.398 { 00:14:20.398 "name": "BaseBdev2", 00:14:20.398 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:20.398 "is_configured": true, 00:14:20.398 "data_offset": 0, 00:14:20.398 "data_size": 65536 00:14:20.398 }, 00:14:20.398 { 00:14:20.398 "name": "BaseBdev3", 00:14:20.398 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:20.398 "is_configured": true, 00:14:20.399 "data_offset": 0, 00:14:20.399 "data_size": 65536 00:14:20.399 } 00:14:20.399 ] 00:14:20.399 }' 00:14:20.399 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:20.399 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:20.399 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:20.399 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:20.399 11:03:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:20.969 [2024-10-29 11:03:26.421580] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:20.969 [2024-10-29 11:03:26.421650] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:20.969 [2024-10-29 11:03:26.421701] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.539 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.539 "name": "raid_bdev1", 00:14:21.539 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:21.539 "strip_size_kb": 64, 00:14:21.539 "state": "online", 00:14:21.539 "raid_level": "raid5f", 00:14:21.539 "superblock": false, 00:14:21.539 "num_base_bdevs": 3, 00:14:21.539 "num_base_bdevs_discovered": 3, 00:14:21.539 "num_base_bdevs_operational": 3, 00:14:21.539 "base_bdevs_list": [ 00:14:21.539 { 00:14:21.539 "name": "spare", 00:14:21.539 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:21.539 "is_configured": true, 00:14:21.539 "data_offset": 0, 00:14:21.539 "data_size": 65536 00:14:21.539 }, 00:14:21.539 { 00:14:21.540 "name": "BaseBdev2", 00:14:21.540 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:21.540 "is_configured": true, 00:14:21.540 "data_offset": 0, 00:14:21.540 "data_size": 65536 00:14:21.540 }, 00:14:21.540 { 00:14:21.540 "name": "BaseBdev3", 00:14:21.540 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:21.540 "is_configured": true, 00:14:21.540 "data_offset": 0, 00:14:21.540 "data_size": 65536 00:14:21.540 } 00:14:21.540 ] 00:14:21.540 }' 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.540 11:03:26 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.540 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.540 "name": "raid_bdev1", 00:14:21.540 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:21.540 "strip_size_kb": 64, 00:14:21.540 "state": "online", 00:14:21.540 "raid_level": "raid5f", 00:14:21.540 "superblock": false, 00:14:21.540 "num_base_bdevs": 3, 00:14:21.540 "num_base_bdevs_discovered": 3, 00:14:21.540 "num_base_bdevs_operational": 3, 00:14:21.540 "base_bdevs_list": [ 00:14:21.540 { 00:14:21.540 "name": "spare", 00:14:21.540 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:21.540 "is_configured": true, 00:14:21.540 "data_offset": 0, 00:14:21.540 "data_size": 65536 00:14:21.540 }, 00:14:21.540 { 00:14:21.540 "name": "BaseBdev2", 00:14:21.540 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:21.540 "is_configured": true, 00:14:21.540 "data_offset": 0, 00:14:21.540 "data_size": 65536 00:14:21.540 }, 00:14:21.540 { 00:14:21.540 "name": "BaseBdev3", 00:14:21.540 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:21.540 "is_configured": true, 00:14:21.540 "data_offset": 0, 00:14:21.540 "data_size": 65536 00:14:21.540 } 00:14:21.540 ] 00:14:21.540 }' 00:14:21.540 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.800 "name": "raid_bdev1", 00:14:21.800 "uuid": "3fb35023-075c-4089-88ba-53a9f3ac84fe", 00:14:21.800 "strip_size_kb": 64, 00:14:21.800 "state": "online", 00:14:21.800 "raid_level": "raid5f", 00:14:21.800 "superblock": false, 00:14:21.800 "num_base_bdevs": 3, 00:14:21.800 "num_base_bdevs_discovered": 3, 00:14:21.800 "num_base_bdevs_operational": 3, 00:14:21.800 "base_bdevs_list": [ 00:14:21.800 { 00:14:21.800 "name": "spare", 00:14:21.800 "uuid": "5392462b-dd10-548e-acfc-901944709e8a", 00:14:21.800 "is_configured": true, 00:14:21.800 "data_offset": 0, 00:14:21.800 "data_size": 65536 00:14:21.800 }, 00:14:21.800 { 00:14:21.800 "name": "BaseBdev2", 00:14:21.800 "uuid": "6dd95683-e52d-5d9b-9dda-df16cb62cee9", 00:14:21.800 "is_configured": true, 00:14:21.800 "data_offset": 0, 00:14:21.800 "data_size": 65536 00:14:21.800 }, 00:14:21.800 { 00:14:21.800 "name": "BaseBdev3", 00:14:21.800 "uuid": "c27d1508-4663-5353-aa60-5c38781390f6", 00:14:21.800 "is_configured": true, 00:14:21.800 "data_offset": 0, 00:14:21.800 "data_size": 65536 00:14:21.800 } 00:14:21.800 ] 00:14:21.800 }' 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.800 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.370 [2024-10-29 11:03:27.577506] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:22.370 [2024-10-29 11:03:27.577589] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:22.370 [2024-10-29 11:03:27.577726] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:22.370 [2024-10-29 11:03:27.577859] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:22.370 [2024-10-29 11:03:27.577929] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.370 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:22.371 /dev/nbd0 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:22.371 1+0 records in 00:14:22.371 1+0 records out 00:14:22.371 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000199249 s, 20.6 MB/s 00:14:22.371 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.631 11:03:27 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:22.631 /dev/nbd1 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:22.631 1+0 records in 00:14:22.631 1+0 records out 00:14:22.631 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000476241 s, 8.6 MB/s 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:22.631 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:22.890 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 92191 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@952 -- # '[' -z 92191 ']' 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # kill -0 92191 00:14:23.150 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@957 -- # uname 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 92191 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 92191' 00:14:23.410 killing process with pid 92191 00:14:23.410 Received shutdown signal, test time was about 60.000000 seconds 00:14:23.410 00:14:23.410 Latency(us) 00:14:23.410 [2024-10-29T11:03:28.908Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:23.410 [2024-10-29T11:03:28.908Z] =================================================================================================================== 00:14:23.410 [2024-10-29T11:03:28.908Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@971 -- # kill 92191 00:14:23.410 [2024-10-29 11:03:28.680411] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:23.410 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@976 -- # wait 92191 00:14:23.410 [2024-10-29 11:03:28.721996] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:23.670 11:03:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:23.670 00:14:23.670 real 0m13.609s 00:14:23.670 user 0m16.948s 00:14:23.670 sys 0m1.977s 00:14:23.670 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:14:23.670 ************************************ 00:14:23.670 END TEST raid5f_rebuild_test 00:14:23.670 11:03:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.671 ************************************ 00:14:23.671 11:03:28 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:14:23.671 11:03:28 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:14:23.671 11:03:28 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:14:23.671 11:03:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:23.671 ************************************ 00:14:23.671 START TEST raid5f_rebuild_test_sb 00:14:23.671 ************************************ 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid5f 3 true false true 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:23.671 11:03:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92614 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92614 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@833 -- # '[' -z 92614 ']' 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:23.671 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:14:23.671 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.671 [2024-10-29 11:03:29.101107] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:14:23.671 [2024-10-29 11:03:29.101348] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92614 ] 00:14:23.671 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:23.671 Zero copy mechanism will not be used. 00:14:23.931 [2024-10-29 11:03:29.274766] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:23.931 [2024-10-29 11:03:29.301562] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:23.931 [2024-10-29 11:03:29.346120] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:23.931 [2024-10-29 11:03:29.346176] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@866 -- # return 0 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.501 BaseBdev1_malloc 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.501 [2024-10-29 11:03:29.937721] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:24.501 [2024-10-29 11:03:29.937790] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:24.501 [2024-10-29 11:03:29.937819] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:24.501 [2024-10-29 11:03:29.937843] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:24.501 [2024-10-29 11:03:29.939995] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:24.501 [2024-10-29 11:03:29.940082] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:24.501 BaseBdev1 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.501 BaseBdev2_malloc 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.501 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.501 [2024-10-29 11:03:29.966607] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:24.501 [2024-10-29 11:03:29.966709] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:24.501 [2024-10-29 11:03:29.966753] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:24.501 [2024-10-29 11:03:29.966785] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:24.502 [2024-10-29 11:03:29.968926] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:24.502 [2024-10-29 11:03:29.969022] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:24.502 BaseBdev2 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.502 BaseBdev3_malloc 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.502 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.502 [2024-10-29 11:03:29.995539] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:24.502 [2024-10-29 11:03:29.995591] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:24.502 [2024-10-29 11:03:29.995633] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:24.502 [2024-10-29 11:03:29.995645] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:24.502 [2024-10-29 11:03:29.997723] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:24.502 [2024-10-29 11:03:29.997815] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:24.761 BaseBdev3 00:14:24.761 11:03:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.761 spare_malloc 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.761 spare_delay 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.761 [2024-10-29 11:03:30.053280] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:24.761 [2024-10-29 11:03:30.053348] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:24.761 [2024-10-29 11:03:30.053406] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:24.761 [2024-10-29 11:03:30.053422] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:24.761 [2024-10-29 11:03:30.056296] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:24.761 [2024-10-29 11:03:30.056350] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:24.761 spare 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.761 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.762 [2024-10-29 11:03:30.065298] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:24.762 [2024-10-29 11:03:30.067202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:24.762 [2024-10-29 11:03:30.067266] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:24.762 [2024-10-29 11:03:30.067445] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:14:24.762 [2024-10-29 11:03:30.067462] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:24.762 [2024-10-29 11:03:30.067730] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:24.762 [2024-10-29 11:03:30.068192] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:14:24.762 [2024-10-29 11:03:30.068205] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:14:24.762 [2024-10-29 11:03:30.068330] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.762 "name": "raid_bdev1", 00:14:24.762 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:24.762 "strip_size_kb": 64, 00:14:24.762 "state": "online", 00:14:24.762 "raid_level": "raid5f", 00:14:24.762 "superblock": true, 00:14:24.762 "num_base_bdevs": 3, 00:14:24.762 "num_base_bdevs_discovered": 3, 00:14:24.762 "num_base_bdevs_operational": 3, 00:14:24.762 "base_bdevs_list": [ 00:14:24.762 { 00:14:24.762 "name": "BaseBdev1", 00:14:24.762 "uuid": "525ce70d-9334-5f1b-b683-b4301eab78fb", 00:14:24.762 "is_configured": true, 00:14:24.762 "data_offset": 2048, 00:14:24.762 "data_size": 63488 00:14:24.762 }, 00:14:24.762 { 00:14:24.762 "name": "BaseBdev2", 00:14:24.762 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:24.762 "is_configured": true, 00:14:24.762 "data_offset": 2048, 00:14:24.762 "data_size": 63488 00:14:24.762 }, 00:14:24.762 { 00:14:24.762 "name": "BaseBdev3", 00:14:24.762 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:24.762 "is_configured": true, 00:14:24.762 "data_offset": 2048, 00:14:24.762 "data_size": 63488 00:14:24.762 } 00:14:24.762 ] 00:14:24.762 }' 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.762 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.022 [2024-10-29 11:03:30.441424] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:25.022 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:25.282 [2024-10-29 11:03:30.704888] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:14:25.282 /dev/nbd0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:25.282 1+0 records in 00:14:25.282 1+0 records out 00:14:25.282 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00038096 s, 10.8 MB/s 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:25.282 11:03:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:14:25.852 496+0 records in 00:14:25.852 496+0 records out 00:14:25.852 65011712 bytes (65 MB, 62 MiB) copied, 0.305482 s, 213 MB/s 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:25.852 [2024-10-29 11:03:31.286255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.852 [2024-10-29 11:03:31.302352] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.852 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.112 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.112 "name": "raid_bdev1", 00:14:26.112 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:26.112 "strip_size_kb": 64, 00:14:26.112 "state": "online", 00:14:26.112 "raid_level": "raid5f", 00:14:26.112 "superblock": true, 00:14:26.112 "num_base_bdevs": 3, 00:14:26.112 "num_base_bdevs_discovered": 2, 00:14:26.112 "num_base_bdevs_operational": 2, 00:14:26.112 "base_bdevs_list": [ 00:14:26.112 { 00:14:26.112 "name": null, 00:14:26.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.112 "is_configured": false, 00:14:26.112 "data_offset": 0, 00:14:26.112 "data_size": 63488 00:14:26.112 }, 00:14:26.112 { 00:14:26.112 "name": "BaseBdev2", 00:14:26.112 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:26.112 "is_configured": true, 00:14:26.112 "data_offset": 2048, 00:14:26.112 "data_size": 63488 00:14:26.112 }, 00:14:26.112 { 00:14:26.112 "name": "BaseBdev3", 00:14:26.112 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:26.112 "is_configured": true, 00:14:26.112 "data_offset": 2048, 00:14:26.112 "data_size": 63488 00:14:26.112 } 00:14:26.112 ] 00:14:26.112 }' 00:14:26.112 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.112 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.372 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:26.372 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.372 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.372 [2024-10-29 11:03:31.749530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:26.372 [2024-10-29 11:03:31.754459] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028de0 00:14:26.372 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.372 11:03:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:26.372 [2024-10-29 11:03:31.756681] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.312 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:27.573 "name": "raid_bdev1", 00:14:27.573 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:27.573 "strip_size_kb": 64, 00:14:27.573 "state": "online", 00:14:27.573 "raid_level": "raid5f", 00:14:27.573 "superblock": true, 00:14:27.573 "num_base_bdevs": 3, 00:14:27.573 "num_base_bdevs_discovered": 3, 00:14:27.573 "num_base_bdevs_operational": 3, 00:14:27.573 "process": { 00:14:27.573 "type": "rebuild", 00:14:27.573 "target": "spare", 00:14:27.573 "progress": { 00:14:27.573 "blocks": 20480, 00:14:27.573 "percent": 16 00:14:27.573 } 00:14:27.573 }, 00:14:27.573 "base_bdevs_list": [ 00:14:27.573 { 00:14:27.573 "name": "spare", 00:14:27.573 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:27.573 "is_configured": true, 00:14:27.573 "data_offset": 2048, 00:14:27.573 "data_size": 63488 00:14:27.573 }, 00:14:27.573 { 00:14:27.573 "name": "BaseBdev2", 00:14:27.573 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:27.573 "is_configured": true, 00:14:27.573 "data_offset": 2048, 00:14:27.573 "data_size": 63488 00:14:27.573 }, 00:14:27.573 { 00:14:27.573 "name": "BaseBdev3", 00:14:27.573 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:27.573 "is_configured": true, 00:14:27.573 "data_offset": 2048, 00:14:27.573 "data_size": 63488 00:14:27.573 } 00:14:27.573 ] 00:14:27.573 }' 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.573 [2024-10-29 11:03:32.900247] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:27.573 [2024-10-29 11:03:32.963808] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:27.573 [2024-10-29 11:03:32.963912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:27.573 [2024-10-29 11:03:32.963930] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:27.573 [2024-10-29 11:03:32.963945] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:27.573 11:03:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:27.573 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.573 "name": "raid_bdev1", 00:14:27.573 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:27.573 "strip_size_kb": 64, 00:14:27.573 "state": "online", 00:14:27.573 "raid_level": "raid5f", 00:14:27.573 "superblock": true, 00:14:27.573 "num_base_bdevs": 3, 00:14:27.573 "num_base_bdevs_discovered": 2, 00:14:27.573 "num_base_bdevs_operational": 2, 00:14:27.573 "base_bdevs_list": [ 00:14:27.573 { 00:14:27.573 "name": null, 00:14:27.573 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.573 "is_configured": false, 00:14:27.573 "data_offset": 0, 00:14:27.573 "data_size": 63488 00:14:27.573 }, 00:14:27.573 { 00:14:27.573 "name": "BaseBdev2", 00:14:27.573 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:27.573 "is_configured": true, 00:14:27.573 "data_offset": 2048, 00:14:27.573 "data_size": 63488 00:14:27.573 }, 00:14:27.573 { 00:14:27.573 "name": "BaseBdev3", 00:14:27.573 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:27.573 "is_configured": true, 00:14:27.573 "data_offset": 2048, 00:14:27.573 "data_size": 63488 00:14:27.573 } 00:14:27.573 ] 00:14:27.573 }' 00:14:27.573 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.573 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:28.144 "name": "raid_bdev1", 00:14:28.144 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:28.144 "strip_size_kb": 64, 00:14:28.144 "state": "online", 00:14:28.144 "raid_level": "raid5f", 00:14:28.144 "superblock": true, 00:14:28.144 "num_base_bdevs": 3, 00:14:28.144 "num_base_bdevs_discovered": 2, 00:14:28.144 "num_base_bdevs_operational": 2, 00:14:28.144 "base_bdevs_list": [ 00:14:28.144 { 00:14:28.144 "name": null, 00:14:28.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:28.144 "is_configured": false, 00:14:28.144 "data_offset": 0, 00:14:28.144 "data_size": 63488 00:14:28.144 }, 00:14:28.144 { 00:14:28.144 "name": "BaseBdev2", 00:14:28.144 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:28.144 "is_configured": true, 00:14:28.144 "data_offset": 2048, 00:14:28.144 "data_size": 63488 00:14:28.144 }, 00:14:28.144 { 00:14:28.144 "name": "BaseBdev3", 00:14:28.144 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:28.144 "is_configured": true, 00:14:28.144 "data_offset": 2048, 00:14:28.144 "data_size": 63488 00:14:28.144 } 00:14:28.144 ] 00:14:28.144 }' 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.144 [2024-10-29 11:03:33.592878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:28.144 [2024-10-29 11:03:33.597413] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028eb0 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.144 11:03:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:28.144 [2024-10-29 11:03:33.599521] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.528 "name": "raid_bdev1", 00:14:29.528 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:29.528 "strip_size_kb": 64, 00:14:29.528 "state": "online", 00:14:29.528 "raid_level": "raid5f", 00:14:29.528 "superblock": true, 00:14:29.528 "num_base_bdevs": 3, 00:14:29.528 "num_base_bdevs_discovered": 3, 00:14:29.528 "num_base_bdevs_operational": 3, 00:14:29.528 "process": { 00:14:29.528 "type": "rebuild", 00:14:29.528 "target": "spare", 00:14:29.528 "progress": { 00:14:29.528 "blocks": 20480, 00:14:29.528 "percent": 16 00:14:29.528 } 00:14:29.528 }, 00:14:29.528 "base_bdevs_list": [ 00:14:29.528 { 00:14:29.528 "name": "spare", 00:14:29.528 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 }, 00:14:29.528 { 00:14:29.528 "name": "BaseBdev2", 00:14:29.528 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 }, 00:14:29.528 { 00:14:29.528 "name": "BaseBdev3", 00:14:29.528 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 } 00:14:29.528 ] 00:14:29.528 }' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:29.528 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=459 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:29.528 "name": "raid_bdev1", 00:14:29.528 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:29.528 "strip_size_kb": 64, 00:14:29.528 "state": "online", 00:14:29.528 "raid_level": "raid5f", 00:14:29.528 "superblock": true, 00:14:29.528 "num_base_bdevs": 3, 00:14:29.528 "num_base_bdevs_discovered": 3, 00:14:29.528 "num_base_bdevs_operational": 3, 00:14:29.528 "process": { 00:14:29.528 "type": "rebuild", 00:14:29.528 "target": "spare", 00:14:29.528 "progress": { 00:14:29.528 "blocks": 22528, 00:14:29.528 "percent": 17 00:14:29.528 } 00:14:29.528 }, 00:14:29.528 "base_bdevs_list": [ 00:14:29.528 { 00:14:29.528 "name": "spare", 00:14:29.528 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 }, 00:14:29.528 { 00:14:29.528 "name": "BaseBdev2", 00:14:29.528 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 }, 00:14:29.528 { 00:14:29.528 "name": "BaseBdev3", 00:14:29.528 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:29.528 "is_configured": true, 00:14:29.528 "data_offset": 2048, 00:14:29.528 "data_size": 63488 00:14:29.528 } 00:14:29.528 ] 00:14:29.528 }' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:29.528 11:03:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:30.468 "name": "raid_bdev1", 00:14:30.468 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:30.468 "strip_size_kb": 64, 00:14:30.468 "state": "online", 00:14:30.468 "raid_level": "raid5f", 00:14:30.468 "superblock": true, 00:14:30.468 "num_base_bdevs": 3, 00:14:30.468 "num_base_bdevs_discovered": 3, 00:14:30.468 "num_base_bdevs_operational": 3, 00:14:30.468 "process": { 00:14:30.468 "type": "rebuild", 00:14:30.468 "target": "spare", 00:14:30.468 "progress": { 00:14:30.468 "blocks": 45056, 00:14:30.468 "percent": 35 00:14:30.468 } 00:14:30.468 }, 00:14:30.468 "base_bdevs_list": [ 00:14:30.468 { 00:14:30.468 "name": "spare", 00:14:30.468 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:30.468 "is_configured": true, 00:14:30.468 "data_offset": 2048, 00:14:30.468 "data_size": 63488 00:14:30.468 }, 00:14:30.468 { 00:14:30.468 "name": "BaseBdev2", 00:14:30.468 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:30.468 "is_configured": true, 00:14:30.468 "data_offset": 2048, 00:14:30.468 "data_size": 63488 00:14:30.468 }, 00:14:30.468 { 00:14:30.468 "name": "BaseBdev3", 00:14:30.468 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:30.468 "is_configured": true, 00:14:30.468 "data_offset": 2048, 00:14:30.468 "data_size": 63488 00:14:30.468 } 00:14:30.468 ] 00:14:30.468 }' 00:14:30.468 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:30.728 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:30.728 11:03:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:30.728 11:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:30.728 11:03:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.668 "name": "raid_bdev1", 00:14:31.668 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:31.668 "strip_size_kb": 64, 00:14:31.668 "state": "online", 00:14:31.668 "raid_level": "raid5f", 00:14:31.668 "superblock": true, 00:14:31.668 "num_base_bdevs": 3, 00:14:31.668 "num_base_bdevs_discovered": 3, 00:14:31.668 "num_base_bdevs_operational": 3, 00:14:31.668 "process": { 00:14:31.668 "type": "rebuild", 00:14:31.668 "target": "spare", 00:14:31.668 "progress": { 00:14:31.668 "blocks": 69632, 00:14:31.668 "percent": 54 00:14:31.668 } 00:14:31.668 }, 00:14:31.668 "base_bdevs_list": [ 00:14:31.668 { 00:14:31.668 "name": "spare", 00:14:31.668 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:31.668 "is_configured": true, 00:14:31.668 "data_offset": 2048, 00:14:31.668 "data_size": 63488 00:14:31.668 }, 00:14:31.668 { 00:14:31.668 "name": "BaseBdev2", 00:14:31.668 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:31.668 "is_configured": true, 00:14:31.668 "data_offset": 2048, 00:14:31.668 "data_size": 63488 00:14:31.668 }, 00:14:31.668 { 00:14:31.668 "name": "BaseBdev3", 00:14:31.668 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:31.668 "is_configured": true, 00:14:31.668 "data_offset": 2048, 00:14:31.668 "data_size": 63488 00:14:31.668 } 00:14:31.668 ] 00:14:31.668 }' 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:31.668 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.928 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.928 11:03:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:32.869 "name": "raid_bdev1", 00:14:32.869 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:32.869 "strip_size_kb": 64, 00:14:32.869 "state": "online", 00:14:32.869 "raid_level": "raid5f", 00:14:32.869 "superblock": true, 00:14:32.869 "num_base_bdevs": 3, 00:14:32.869 "num_base_bdevs_discovered": 3, 00:14:32.869 "num_base_bdevs_operational": 3, 00:14:32.869 "process": { 00:14:32.869 "type": "rebuild", 00:14:32.869 "target": "spare", 00:14:32.869 "progress": { 00:14:32.869 "blocks": 92160, 00:14:32.869 "percent": 72 00:14:32.869 } 00:14:32.869 }, 00:14:32.869 "base_bdevs_list": [ 00:14:32.869 { 00:14:32.869 "name": "spare", 00:14:32.869 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:32.869 "is_configured": true, 00:14:32.869 "data_offset": 2048, 00:14:32.869 "data_size": 63488 00:14:32.869 }, 00:14:32.869 { 00:14:32.869 "name": "BaseBdev2", 00:14:32.869 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:32.869 "is_configured": true, 00:14:32.869 "data_offset": 2048, 00:14:32.869 "data_size": 63488 00:14:32.869 }, 00:14:32.869 { 00:14:32.869 "name": "BaseBdev3", 00:14:32.869 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:32.869 "is_configured": true, 00:14:32.869 "data_offset": 2048, 00:14:32.869 "data_size": 63488 00:14:32.869 } 00:14:32.869 ] 00:14:32.869 }' 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:32.869 11:03:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:34.251 "name": "raid_bdev1", 00:14:34.251 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:34.251 "strip_size_kb": 64, 00:14:34.251 "state": "online", 00:14:34.251 "raid_level": "raid5f", 00:14:34.251 "superblock": true, 00:14:34.251 "num_base_bdevs": 3, 00:14:34.251 "num_base_bdevs_discovered": 3, 00:14:34.251 "num_base_bdevs_operational": 3, 00:14:34.251 "process": { 00:14:34.251 "type": "rebuild", 00:14:34.251 "target": "spare", 00:14:34.251 "progress": { 00:14:34.251 "blocks": 116736, 00:14:34.251 "percent": 91 00:14:34.251 } 00:14:34.251 }, 00:14:34.251 "base_bdevs_list": [ 00:14:34.251 { 00:14:34.251 "name": "spare", 00:14:34.251 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:34.251 "is_configured": true, 00:14:34.251 "data_offset": 2048, 00:14:34.251 "data_size": 63488 00:14:34.251 }, 00:14:34.251 { 00:14:34.251 "name": "BaseBdev2", 00:14:34.251 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:34.251 "is_configured": true, 00:14:34.251 "data_offset": 2048, 00:14:34.251 "data_size": 63488 00:14:34.251 }, 00:14:34.251 { 00:14:34.251 "name": "BaseBdev3", 00:14:34.251 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:34.251 "is_configured": true, 00:14:34.251 "data_offset": 2048, 00:14:34.251 "data_size": 63488 00:14:34.251 } 00:14:34.251 ] 00:14:34.251 }' 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:34.251 11:03:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:34.511 [2024-10-29 11:03:39.833203] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:34.511 [2024-10-29 11:03:39.833325] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:34.511 [2024-10-29 11:03:39.833493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.080 "name": "raid_bdev1", 00:14:35.080 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:35.080 "strip_size_kb": 64, 00:14:35.080 "state": "online", 00:14:35.080 "raid_level": "raid5f", 00:14:35.080 "superblock": true, 00:14:35.080 "num_base_bdevs": 3, 00:14:35.080 "num_base_bdevs_discovered": 3, 00:14:35.080 "num_base_bdevs_operational": 3, 00:14:35.080 "base_bdevs_list": [ 00:14:35.080 { 00:14:35.080 "name": "spare", 00:14:35.080 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:35.080 "is_configured": true, 00:14:35.080 "data_offset": 2048, 00:14:35.080 "data_size": 63488 00:14:35.080 }, 00:14:35.080 { 00:14:35.080 "name": "BaseBdev2", 00:14:35.080 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:35.080 "is_configured": true, 00:14:35.080 "data_offset": 2048, 00:14:35.080 "data_size": 63488 00:14:35.080 }, 00:14:35.080 { 00:14:35.080 "name": "BaseBdev3", 00:14:35.080 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:35.080 "is_configured": true, 00:14:35.080 "data_offset": 2048, 00:14:35.080 "data_size": 63488 00:14:35.080 } 00:14:35.080 ] 00:14:35.080 }' 00:14:35.080 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.341 "name": "raid_bdev1", 00:14:35.341 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:35.341 "strip_size_kb": 64, 00:14:35.341 "state": "online", 00:14:35.341 "raid_level": "raid5f", 00:14:35.341 "superblock": true, 00:14:35.341 "num_base_bdevs": 3, 00:14:35.341 "num_base_bdevs_discovered": 3, 00:14:35.341 "num_base_bdevs_operational": 3, 00:14:35.341 "base_bdevs_list": [ 00:14:35.341 { 00:14:35.341 "name": "spare", 00:14:35.341 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 }, 00:14:35.341 { 00:14:35.341 "name": "BaseBdev2", 00:14:35.341 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 }, 00:14:35.341 { 00:14:35.341 "name": "BaseBdev3", 00:14:35.341 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 } 00:14:35.341 ] 00:14:35.341 }' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.341 "name": "raid_bdev1", 00:14:35.341 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:35.341 "strip_size_kb": 64, 00:14:35.341 "state": "online", 00:14:35.341 "raid_level": "raid5f", 00:14:35.341 "superblock": true, 00:14:35.341 "num_base_bdevs": 3, 00:14:35.341 "num_base_bdevs_discovered": 3, 00:14:35.341 "num_base_bdevs_operational": 3, 00:14:35.341 "base_bdevs_list": [ 00:14:35.341 { 00:14:35.341 "name": "spare", 00:14:35.341 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 }, 00:14:35.341 { 00:14:35.341 "name": "BaseBdev2", 00:14:35.341 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 }, 00:14:35.341 { 00:14:35.341 "name": "BaseBdev3", 00:14:35.341 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:35.341 "is_configured": true, 00:14:35.341 "data_offset": 2048, 00:14:35.341 "data_size": 63488 00:14:35.341 } 00:14:35.341 ] 00:14:35.341 }' 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.341 11:03:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.912 [2024-10-29 11:03:41.268659] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:35.912 [2024-10-29 11:03:41.268743] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:35.912 [2024-10-29 11:03:41.268865] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:35.912 [2024-10-29 11:03:41.268950] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:35.912 [2024-10-29 11:03:41.268971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:35.912 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:36.172 /dev/nbd0 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:36.172 1+0 records in 00:14:36.172 1+0 records out 00:14:36.172 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000594149 s, 6.9 MB/s 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:36.172 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:36.433 /dev/nbd1 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:36.433 1+0 records in 00:14:36.433 1+0 records out 00:14:36.433 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000455003 s, 9.0 MB/s 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:36.433 11:03:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:36.693 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.954 [2024-10-29 11:03:42.338705] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:36.954 [2024-10-29 11:03:42.338770] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:36.954 [2024-10-29 11:03:42.338796] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:36.954 [2024-10-29 11:03:42.338806] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:36.954 [2024-10-29 11:03:42.340976] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:36.954 [2024-10-29 11:03:42.341020] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:36.954 [2024-10-29 11:03:42.341128] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:36.954 [2024-10-29 11:03:42.341166] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:36.954 [2024-10-29 11:03:42.341273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:36.954 [2024-10-29 11:03:42.341393] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:36.954 spare 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.954 [2024-10-29 11:03:42.441342] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:14:36.954 [2024-10-29 11:03:42.441380] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:36.954 [2024-10-29 11:03:42.441615] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047560 00:14:36.954 [2024-10-29 11:03:42.442106] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:14:36.954 [2024-10-29 11:03:42.442133] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:14:36.954 [2024-10-29 11:03:42.442259] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.954 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.955 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.215 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.215 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.215 "name": "raid_bdev1", 00:14:37.215 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:37.215 "strip_size_kb": 64, 00:14:37.215 "state": "online", 00:14:37.215 "raid_level": "raid5f", 00:14:37.215 "superblock": true, 00:14:37.215 "num_base_bdevs": 3, 00:14:37.215 "num_base_bdevs_discovered": 3, 00:14:37.215 "num_base_bdevs_operational": 3, 00:14:37.215 "base_bdevs_list": [ 00:14:37.215 { 00:14:37.215 "name": "spare", 00:14:37.215 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:37.215 "is_configured": true, 00:14:37.215 "data_offset": 2048, 00:14:37.215 "data_size": 63488 00:14:37.215 }, 00:14:37.215 { 00:14:37.215 "name": "BaseBdev2", 00:14:37.215 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:37.215 "is_configured": true, 00:14:37.215 "data_offset": 2048, 00:14:37.215 "data_size": 63488 00:14:37.215 }, 00:14:37.215 { 00:14:37.215 "name": "BaseBdev3", 00:14:37.215 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:37.215 "is_configured": true, 00:14:37.215 "data_offset": 2048, 00:14:37.215 "data_size": 63488 00:14:37.215 } 00:14:37.215 ] 00:14:37.215 }' 00:14:37.215 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.215 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.475 "name": "raid_bdev1", 00:14:37.475 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:37.475 "strip_size_kb": 64, 00:14:37.475 "state": "online", 00:14:37.475 "raid_level": "raid5f", 00:14:37.475 "superblock": true, 00:14:37.475 "num_base_bdevs": 3, 00:14:37.475 "num_base_bdevs_discovered": 3, 00:14:37.475 "num_base_bdevs_operational": 3, 00:14:37.475 "base_bdevs_list": [ 00:14:37.475 { 00:14:37.475 "name": "spare", 00:14:37.475 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:37.475 "is_configured": true, 00:14:37.475 "data_offset": 2048, 00:14:37.475 "data_size": 63488 00:14:37.475 }, 00:14:37.475 { 00:14:37.475 "name": "BaseBdev2", 00:14:37.475 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:37.475 "is_configured": true, 00:14:37.475 "data_offset": 2048, 00:14:37.475 "data_size": 63488 00:14:37.475 }, 00:14:37.475 { 00:14:37.475 "name": "BaseBdev3", 00:14:37.475 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:37.475 "is_configured": true, 00:14:37.475 "data_offset": 2048, 00:14:37.475 "data_size": 63488 00:14:37.475 } 00:14:37.475 ] 00:14:37.475 }' 00:14:37.475 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.735 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:37.735 11:03:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.735 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:37.735 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.736 [2024-10-29 11:03:43.074476] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.736 "name": "raid_bdev1", 00:14:37.736 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:37.736 "strip_size_kb": 64, 00:14:37.736 "state": "online", 00:14:37.736 "raid_level": "raid5f", 00:14:37.736 "superblock": true, 00:14:37.736 "num_base_bdevs": 3, 00:14:37.736 "num_base_bdevs_discovered": 2, 00:14:37.736 "num_base_bdevs_operational": 2, 00:14:37.736 "base_bdevs_list": [ 00:14:37.736 { 00:14:37.736 "name": null, 00:14:37.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.736 "is_configured": false, 00:14:37.736 "data_offset": 0, 00:14:37.736 "data_size": 63488 00:14:37.736 }, 00:14:37.736 { 00:14:37.736 "name": "BaseBdev2", 00:14:37.736 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:37.736 "is_configured": true, 00:14:37.736 "data_offset": 2048, 00:14:37.736 "data_size": 63488 00:14:37.736 }, 00:14:37.736 { 00:14:37.736 "name": "BaseBdev3", 00:14:37.736 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:37.736 "is_configured": true, 00:14:37.736 "data_offset": 2048, 00:14:37.736 "data_size": 63488 00:14:37.736 } 00:14:37.736 ] 00:14:37.736 }' 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.736 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.306 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:38.306 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.306 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.306 [2024-10-29 11:03:43.525645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:38.306 [2024-10-29 11:03:43.525811] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:38.306 [2024-10-29 11:03:43.525835] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:38.306 [2024-10-29 11:03:43.525876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:38.306 [2024-10-29 11:03:43.530448] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047630 00:14:38.306 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:38.306 11:03:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:38.306 [2024-10-29 11:03:43.532631] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.247 "name": "raid_bdev1", 00:14:39.247 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:39.247 "strip_size_kb": 64, 00:14:39.247 "state": "online", 00:14:39.247 "raid_level": "raid5f", 00:14:39.247 "superblock": true, 00:14:39.247 "num_base_bdevs": 3, 00:14:39.247 "num_base_bdevs_discovered": 3, 00:14:39.247 "num_base_bdevs_operational": 3, 00:14:39.247 "process": { 00:14:39.247 "type": "rebuild", 00:14:39.247 "target": "spare", 00:14:39.247 "progress": { 00:14:39.247 "blocks": 20480, 00:14:39.247 "percent": 16 00:14:39.247 } 00:14:39.247 }, 00:14:39.247 "base_bdevs_list": [ 00:14:39.247 { 00:14:39.247 "name": "spare", 00:14:39.247 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 2048, 00:14:39.247 "data_size": 63488 00:14:39.247 }, 00:14:39.247 { 00:14:39.247 "name": "BaseBdev2", 00:14:39.247 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 2048, 00:14:39.247 "data_size": 63488 00:14:39.247 }, 00:14:39.247 { 00:14:39.247 "name": "BaseBdev3", 00:14:39.247 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:39.247 "is_configured": true, 00:14:39.247 "data_offset": 2048, 00:14:39.247 "data_size": 63488 00:14:39.247 } 00:14:39.247 ] 00:14:39.247 }' 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.247 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.247 [2024-10-29 11:03:44.688647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:39.247 [2024-10-29 11:03:44.739410] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:39.247 [2024-10-29 11:03:44.739483] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:39.247 [2024-10-29 11:03:44.739523] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:39.247 [2024-10-29 11:03:44.739532] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.508 "name": "raid_bdev1", 00:14:39.508 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:39.508 "strip_size_kb": 64, 00:14:39.508 "state": "online", 00:14:39.508 "raid_level": "raid5f", 00:14:39.508 "superblock": true, 00:14:39.508 "num_base_bdevs": 3, 00:14:39.508 "num_base_bdevs_discovered": 2, 00:14:39.508 "num_base_bdevs_operational": 2, 00:14:39.508 "base_bdevs_list": [ 00:14:39.508 { 00:14:39.508 "name": null, 00:14:39.508 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.508 "is_configured": false, 00:14:39.508 "data_offset": 0, 00:14:39.508 "data_size": 63488 00:14:39.508 }, 00:14:39.508 { 00:14:39.508 "name": "BaseBdev2", 00:14:39.508 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:39.508 "is_configured": true, 00:14:39.508 "data_offset": 2048, 00:14:39.508 "data_size": 63488 00:14:39.508 }, 00:14:39.508 { 00:14:39.508 "name": "BaseBdev3", 00:14:39.508 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:39.508 "is_configured": true, 00:14:39.508 "data_offset": 2048, 00:14:39.508 "data_size": 63488 00:14:39.508 } 00:14:39.508 ] 00:14:39.508 }' 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.508 11:03:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.769 11:03:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:39.769 11:03:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.769 11:03:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.769 [2024-10-29 11:03:45.224191] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:39.769 [2024-10-29 11:03:45.224256] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:39.769 [2024-10-29 11:03:45.224281] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:14:39.769 [2024-10-29 11:03:45.224293] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:39.769 [2024-10-29 11:03:45.224820] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:39.769 [2024-10-29 11:03:45.224852] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:39.769 [2024-10-29 11:03:45.224953] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:39.769 [2024-10-29 11:03:45.224975] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:39.769 [2024-10-29 11:03:45.224988] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:39.769 [2024-10-29 11:03:45.225011] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:39.769 [2024-10-29 11:03:45.229109] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047700 00:14:39.769 spare 00:14:39.769 11:03:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.769 11:03:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:39.769 [2024-10-29 11:03:45.231308] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.153 "name": "raid_bdev1", 00:14:41.153 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:41.153 "strip_size_kb": 64, 00:14:41.153 "state": "online", 00:14:41.153 "raid_level": "raid5f", 00:14:41.153 "superblock": true, 00:14:41.153 "num_base_bdevs": 3, 00:14:41.153 "num_base_bdevs_discovered": 3, 00:14:41.153 "num_base_bdevs_operational": 3, 00:14:41.153 "process": { 00:14:41.153 "type": "rebuild", 00:14:41.153 "target": "spare", 00:14:41.153 "progress": { 00:14:41.153 "blocks": 20480, 00:14:41.153 "percent": 16 00:14:41.153 } 00:14:41.153 }, 00:14:41.153 "base_bdevs_list": [ 00:14:41.153 { 00:14:41.153 "name": "spare", 00:14:41.153 "uuid": "cb3b8192-2052-5272-93ba-19cc4c8977ed", 00:14:41.153 "is_configured": true, 00:14:41.153 "data_offset": 2048, 00:14:41.153 "data_size": 63488 00:14:41.153 }, 00:14:41.153 { 00:14:41.153 "name": "BaseBdev2", 00:14:41.153 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:41.153 "is_configured": true, 00:14:41.153 "data_offset": 2048, 00:14:41.153 "data_size": 63488 00:14:41.153 }, 00:14:41.153 { 00:14:41.153 "name": "BaseBdev3", 00:14:41.153 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:41.153 "is_configured": true, 00:14:41.153 "data_offset": 2048, 00:14:41.153 "data_size": 63488 00:14:41.153 } 00:14:41.153 ] 00:14:41.153 }' 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:41.153 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.154 [2024-10-29 11:03:46.412063] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:41.154 [2024-10-29 11:03:46.438126] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:41.154 [2024-10-29 11:03:46.438211] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:41.154 [2024-10-29 11:03:46.438229] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:41.154 [2024-10-29 11:03:46.438244] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:41.154 "name": "raid_bdev1", 00:14:41.154 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:41.154 "strip_size_kb": 64, 00:14:41.154 "state": "online", 00:14:41.154 "raid_level": "raid5f", 00:14:41.154 "superblock": true, 00:14:41.154 "num_base_bdevs": 3, 00:14:41.154 "num_base_bdevs_discovered": 2, 00:14:41.154 "num_base_bdevs_operational": 2, 00:14:41.154 "base_bdevs_list": [ 00:14:41.154 { 00:14:41.154 "name": null, 00:14:41.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.154 "is_configured": false, 00:14:41.154 "data_offset": 0, 00:14:41.154 "data_size": 63488 00:14:41.154 }, 00:14:41.154 { 00:14:41.154 "name": "BaseBdev2", 00:14:41.154 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:41.154 "is_configured": true, 00:14:41.154 "data_offset": 2048, 00:14:41.154 "data_size": 63488 00:14:41.154 }, 00:14:41.154 { 00:14:41.154 "name": "BaseBdev3", 00:14:41.154 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:41.154 "is_configured": true, 00:14:41.154 "data_offset": 2048, 00:14:41.154 "data_size": 63488 00:14:41.154 } 00:14:41.154 ] 00:14:41.154 }' 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:41.154 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.724 "name": "raid_bdev1", 00:14:41.724 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:41.724 "strip_size_kb": 64, 00:14:41.724 "state": "online", 00:14:41.724 "raid_level": "raid5f", 00:14:41.724 "superblock": true, 00:14:41.724 "num_base_bdevs": 3, 00:14:41.724 "num_base_bdevs_discovered": 2, 00:14:41.724 "num_base_bdevs_operational": 2, 00:14:41.724 "base_bdevs_list": [ 00:14:41.724 { 00:14:41.724 "name": null, 00:14:41.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:41.724 "is_configured": false, 00:14:41.724 "data_offset": 0, 00:14:41.724 "data_size": 63488 00:14:41.724 }, 00:14:41.724 { 00:14:41.724 "name": "BaseBdev2", 00:14:41.724 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:41.724 "is_configured": true, 00:14:41.724 "data_offset": 2048, 00:14:41.724 "data_size": 63488 00:14:41.724 }, 00:14:41.724 { 00:14:41.724 "name": "BaseBdev3", 00:14:41.724 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:41.724 "is_configured": true, 00:14:41.724 "data_offset": 2048, 00:14:41.724 "data_size": 63488 00:14:41.724 } 00:14:41.724 ] 00:14:41.724 }' 00:14:41.724 11:03:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.724 [2024-10-29 11:03:47.074733] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:41.724 [2024-10-29 11:03:47.074792] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:41.724 [2024-10-29 11:03:47.074817] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:14:41.724 [2024-10-29 11:03:47.074830] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:41.724 [2024-10-29 11:03:47.075252] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:41.724 [2024-10-29 11:03:47.075286] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:41.724 [2024-10-29 11:03:47.075375] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:41.724 [2024-10-29 11:03:47.075394] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:41.724 [2024-10-29 11:03:47.075403] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:41.724 [2024-10-29 11:03:47.075417] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:41.724 BaseBdev1 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.724 11:03:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.666 "name": "raid_bdev1", 00:14:42.666 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:42.666 "strip_size_kb": 64, 00:14:42.666 "state": "online", 00:14:42.666 "raid_level": "raid5f", 00:14:42.666 "superblock": true, 00:14:42.666 "num_base_bdevs": 3, 00:14:42.666 "num_base_bdevs_discovered": 2, 00:14:42.666 "num_base_bdevs_operational": 2, 00:14:42.666 "base_bdevs_list": [ 00:14:42.666 { 00:14:42.666 "name": null, 00:14:42.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.666 "is_configured": false, 00:14:42.666 "data_offset": 0, 00:14:42.666 "data_size": 63488 00:14:42.666 }, 00:14:42.666 { 00:14:42.666 "name": "BaseBdev2", 00:14:42.666 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:42.666 "is_configured": true, 00:14:42.666 "data_offset": 2048, 00:14:42.666 "data_size": 63488 00:14:42.666 }, 00:14:42.666 { 00:14:42.666 "name": "BaseBdev3", 00:14:42.666 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:42.666 "is_configured": true, 00:14:42.666 "data_offset": 2048, 00:14:42.666 "data_size": 63488 00:14:42.666 } 00:14:42.666 ] 00:14:42.666 }' 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.666 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:43.237 "name": "raid_bdev1", 00:14:43.237 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:43.237 "strip_size_kb": 64, 00:14:43.237 "state": "online", 00:14:43.237 "raid_level": "raid5f", 00:14:43.237 "superblock": true, 00:14:43.237 "num_base_bdevs": 3, 00:14:43.237 "num_base_bdevs_discovered": 2, 00:14:43.237 "num_base_bdevs_operational": 2, 00:14:43.237 "base_bdevs_list": [ 00:14:43.237 { 00:14:43.237 "name": null, 00:14:43.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.237 "is_configured": false, 00:14:43.237 "data_offset": 0, 00:14:43.237 "data_size": 63488 00:14:43.237 }, 00:14:43.237 { 00:14:43.237 "name": "BaseBdev2", 00:14:43.237 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:43.237 "is_configured": true, 00:14:43.237 "data_offset": 2048, 00:14:43.237 "data_size": 63488 00:14:43.237 }, 00:14:43.237 { 00:14:43.237 "name": "BaseBdev3", 00:14:43.237 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:43.237 "is_configured": true, 00:14:43.237 "data_offset": 2048, 00:14:43.237 "data_size": 63488 00:14:43.237 } 00:14:43.237 ] 00:14:43.237 }' 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.237 [2024-10-29 11:03:48.704001] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:43.237 [2024-10-29 11:03:48.704193] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:43.237 [2024-10-29 11:03:48.704207] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:43.237 request: 00:14:43.237 { 00:14:43.237 "base_bdev": "BaseBdev1", 00:14:43.237 "raid_bdev": "raid_bdev1", 00:14:43.237 "method": "bdev_raid_add_base_bdev", 00:14:43.237 "req_id": 1 00:14:43.237 } 00:14:43.237 Got JSON-RPC error response 00:14:43.237 response: 00:14:43.237 { 00:14:43.237 "code": -22, 00:14:43.237 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:43.237 } 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:43.237 11:03:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.619 "name": "raid_bdev1", 00:14:44.619 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:44.619 "strip_size_kb": 64, 00:14:44.619 "state": "online", 00:14:44.619 "raid_level": "raid5f", 00:14:44.619 "superblock": true, 00:14:44.619 "num_base_bdevs": 3, 00:14:44.619 "num_base_bdevs_discovered": 2, 00:14:44.619 "num_base_bdevs_operational": 2, 00:14:44.619 "base_bdevs_list": [ 00:14:44.619 { 00:14:44.619 "name": null, 00:14:44.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.619 "is_configured": false, 00:14:44.619 "data_offset": 0, 00:14:44.619 "data_size": 63488 00:14:44.619 }, 00:14:44.619 { 00:14:44.619 "name": "BaseBdev2", 00:14:44.619 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:44.619 "is_configured": true, 00:14:44.619 "data_offset": 2048, 00:14:44.619 "data_size": 63488 00:14:44.619 }, 00:14:44.619 { 00:14:44.619 "name": "BaseBdev3", 00:14:44.619 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:44.619 "is_configured": true, 00:14:44.619 "data_offset": 2048, 00:14:44.619 "data_size": 63488 00:14:44.619 } 00:14:44.619 ] 00:14:44.619 }' 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.619 11:03:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.879 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:44.879 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:44.879 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:44.879 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:44.879 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:44.880 "name": "raid_bdev1", 00:14:44.880 "uuid": "d05ce9a4-03e6-48b7-95fb-44dd74d3297c", 00:14:44.880 "strip_size_kb": 64, 00:14:44.880 "state": "online", 00:14:44.880 "raid_level": "raid5f", 00:14:44.880 "superblock": true, 00:14:44.880 "num_base_bdevs": 3, 00:14:44.880 "num_base_bdevs_discovered": 2, 00:14:44.880 "num_base_bdevs_operational": 2, 00:14:44.880 "base_bdevs_list": [ 00:14:44.880 { 00:14:44.880 "name": null, 00:14:44.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:44.880 "is_configured": false, 00:14:44.880 "data_offset": 0, 00:14:44.880 "data_size": 63488 00:14:44.880 }, 00:14:44.880 { 00:14:44.880 "name": "BaseBdev2", 00:14:44.880 "uuid": "9988291c-5e58-524b-a508-75c452894551", 00:14:44.880 "is_configured": true, 00:14:44.880 "data_offset": 2048, 00:14:44.880 "data_size": 63488 00:14:44.880 }, 00:14:44.880 { 00:14:44.880 "name": "BaseBdev3", 00:14:44.880 "uuid": "184f4e84-ae7c-5312-b258-370fa150619e", 00:14:44.880 "is_configured": true, 00:14:44.880 "data_offset": 2048, 00:14:44.880 "data_size": 63488 00:14:44.880 } 00:14:44.880 ] 00:14:44.880 }' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92614 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@952 -- # '[' -z 92614 ']' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # kill -0 92614 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@957 -- # uname 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 92614 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 92614' 00:14:44.880 killing process with pid 92614 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@971 -- # kill 92614 00:14:44.880 Received shutdown signal, test time was about 60.000000 seconds 00:14:44.880 00:14:44.880 Latency(us) 00:14:44.880 [2024-10-29T11:03:50.378Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:44.880 [2024-10-29T11:03:50.378Z] =================================================================================================================== 00:14:44.880 [2024-10-29T11:03:50.378Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:44.880 [2024-10-29 11:03:50.376304] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:44.880 [2024-10-29 11:03:50.376472] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:44.880 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@976 -- # wait 92614 00:14:44.880 [2024-10-29 11:03:50.376558] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:44.880 [2024-10-29 11:03:50.376569] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:14:45.140 [2024-10-29 11:03:50.417638] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:45.140 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:45.140 00:14:45.140 real 0m21.621s 00:14:45.140 user 0m28.107s 00:14:45.140 sys 0m2.826s 00:14:45.140 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:14:45.140 11:03:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.140 ************************************ 00:14:45.140 END TEST raid5f_rebuild_test_sb 00:14:45.140 ************************************ 00:14:45.401 11:03:50 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:45.401 11:03:50 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:45.401 11:03:50 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:14:45.401 11:03:50 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:14:45.401 11:03:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:45.401 ************************************ 00:14:45.401 START TEST raid5f_state_function_test 00:14:45.401 ************************************ 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1127 -- # raid_state_function_test raid5f 4 false 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=93344 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93344' 00:14:45.401 Process raid pid: 93344 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 93344 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@833 -- # '[' -z 93344 ']' 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:14:45.401 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:14:45.401 11:03:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.401 [2024-10-29 11:03:50.806398] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:14:45.401 [2024-10-29 11:03:50.806542] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:45.662 [2024-10-29 11:03:50.981203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:45.662 [2024-10-29 11:03:51.008459] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:45.662 [2024-10-29 11:03:51.053471] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:45.662 [2024-10-29 11:03:51.053523] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@866 -- # return 0 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.232 [2024-10-29 11:03:51.624113] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:46.232 [2024-10-29 11:03:51.624179] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:46.232 [2024-10-29 11:03:51.624190] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:46.232 [2024-10-29 11:03:51.624201] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:46.232 [2024-10-29 11:03:51.624209] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:46.232 [2024-10-29 11:03:51.624221] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:46.232 [2024-10-29 11:03:51.624229] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:46.232 [2024-10-29 11:03:51.624239] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.232 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.232 "name": "Existed_Raid", 00:14:46.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.232 "strip_size_kb": 64, 00:14:46.232 "state": "configuring", 00:14:46.232 "raid_level": "raid5f", 00:14:46.232 "superblock": false, 00:14:46.232 "num_base_bdevs": 4, 00:14:46.232 "num_base_bdevs_discovered": 0, 00:14:46.232 "num_base_bdevs_operational": 4, 00:14:46.232 "base_bdevs_list": [ 00:14:46.232 { 00:14:46.232 "name": "BaseBdev1", 00:14:46.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.232 "is_configured": false, 00:14:46.232 "data_offset": 0, 00:14:46.232 "data_size": 0 00:14:46.232 }, 00:14:46.232 { 00:14:46.232 "name": "BaseBdev2", 00:14:46.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.232 "is_configured": false, 00:14:46.232 "data_offset": 0, 00:14:46.232 "data_size": 0 00:14:46.232 }, 00:14:46.232 { 00:14:46.232 "name": "BaseBdev3", 00:14:46.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.233 "is_configured": false, 00:14:46.233 "data_offset": 0, 00:14:46.233 "data_size": 0 00:14:46.233 }, 00:14:46.233 { 00:14:46.233 "name": "BaseBdev4", 00:14:46.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.233 "is_configured": false, 00:14:46.233 "data_offset": 0, 00:14:46.233 "data_size": 0 00:14:46.233 } 00:14:46.233 ] 00:14:46.233 }' 00:14:46.233 11:03:51 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.233 11:03:51 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 [2024-10-29 11:03:52.111309] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:46.803 [2024-10-29 11:03:52.111367] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 [2024-10-29 11:03:52.123302] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:46.803 [2024-10-29 11:03:52.123368] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:46.803 [2024-10-29 11:03:52.123388] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:46.803 [2024-10-29 11:03:52.123400] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:46.803 [2024-10-29 11:03:52.123408] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:46.803 [2024-10-29 11:03:52.123418] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:46.803 [2024-10-29 11:03:52.123426] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:46.803 [2024-10-29 11:03:52.123437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 [2024-10-29 11:03:52.144625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:46.803 BaseBdev1 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.803 [ 00:14:46.803 { 00:14:46.803 "name": "BaseBdev1", 00:14:46.803 "aliases": [ 00:14:46.803 "ba01531a-7532-4fb6-bd16-ff0890af2ee2" 00:14:46.803 ], 00:14:46.803 "product_name": "Malloc disk", 00:14:46.803 "block_size": 512, 00:14:46.803 "num_blocks": 65536, 00:14:46.803 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:46.803 "assigned_rate_limits": { 00:14:46.803 "rw_ios_per_sec": 0, 00:14:46.803 "rw_mbytes_per_sec": 0, 00:14:46.803 "r_mbytes_per_sec": 0, 00:14:46.803 "w_mbytes_per_sec": 0 00:14:46.803 }, 00:14:46.803 "claimed": true, 00:14:46.803 "claim_type": "exclusive_write", 00:14:46.803 "zoned": false, 00:14:46.803 "supported_io_types": { 00:14:46.803 "read": true, 00:14:46.803 "write": true, 00:14:46.803 "unmap": true, 00:14:46.803 "flush": true, 00:14:46.803 "reset": true, 00:14:46.803 "nvme_admin": false, 00:14:46.803 "nvme_io": false, 00:14:46.803 "nvme_io_md": false, 00:14:46.803 "write_zeroes": true, 00:14:46.803 "zcopy": true, 00:14:46.803 "get_zone_info": false, 00:14:46.803 "zone_management": false, 00:14:46.803 "zone_append": false, 00:14:46.803 "compare": false, 00:14:46.803 "compare_and_write": false, 00:14:46.803 "abort": true, 00:14:46.803 "seek_hole": false, 00:14:46.803 "seek_data": false, 00:14:46.803 "copy": true, 00:14:46.803 "nvme_iov_md": false 00:14:46.803 }, 00:14:46.803 "memory_domains": [ 00:14:46.803 { 00:14:46.803 "dma_device_id": "system", 00:14:46.803 "dma_device_type": 1 00:14:46.803 }, 00:14:46.803 { 00:14:46.803 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:46.803 "dma_device_type": 2 00:14:46.803 } 00:14:46.803 ], 00:14:46.803 "driver_specific": {} 00:14:46.803 } 00:14:46.803 ] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.803 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.804 "name": "Existed_Raid", 00:14:46.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.804 "strip_size_kb": 64, 00:14:46.804 "state": "configuring", 00:14:46.804 "raid_level": "raid5f", 00:14:46.804 "superblock": false, 00:14:46.804 "num_base_bdevs": 4, 00:14:46.804 "num_base_bdevs_discovered": 1, 00:14:46.804 "num_base_bdevs_operational": 4, 00:14:46.804 "base_bdevs_list": [ 00:14:46.804 { 00:14:46.804 "name": "BaseBdev1", 00:14:46.804 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:46.804 "is_configured": true, 00:14:46.804 "data_offset": 0, 00:14:46.804 "data_size": 65536 00:14:46.804 }, 00:14:46.804 { 00:14:46.804 "name": "BaseBdev2", 00:14:46.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.804 "is_configured": false, 00:14:46.804 "data_offset": 0, 00:14:46.804 "data_size": 0 00:14:46.804 }, 00:14:46.804 { 00:14:46.804 "name": "BaseBdev3", 00:14:46.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.804 "is_configured": false, 00:14:46.804 "data_offset": 0, 00:14:46.804 "data_size": 0 00:14:46.804 }, 00:14:46.804 { 00:14:46.804 "name": "BaseBdev4", 00:14:46.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.804 "is_configured": false, 00:14:46.804 "data_offset": 0, 00:14:46.804 "data_size": 0 00:14:46.804 } 00:14:46.804 ] 00:14:46.804 }' 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.804 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.064 [2024-10-29 11:03:52.548001] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:47.064 [2024-10-29 11:03:52.548049] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.064 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.064 [2024-10-29 11:03:52.560038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:47.064 [2024-10-29 11:03:52.561947] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:47.064 [2024-10-29 11:03:52.561992] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:47.064 [2024-10-29 11:03:52.562003] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:47.064 [2024-10-29 11:03:52.562013] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:47.064 [2024-10-29 11:03:52.562037] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:47.064 [2024-10-29 11:03:52.562047] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:47.324 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.324 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:47.324 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.325 "name": "Existed_Raid", 00:14:47.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.325 "strip_size_kb": 64, 00:14:47.325 "state": "configuring", 00:14:47.325 "raid_level": "raid5f", 00:14:47.325 "superblock": false, 00:14:47.325 "num_base_bdevs": 4, 00:14:47.325 "num_base_bdevs_discovered": 1, 00:14:47.325 "num_base_bdevs_operational": 4, 00:14:47.325 "base_bdevs_list": [ 00:14:47.325 { 00:14:47.325 "name": "BaseBdev1", 00:14:47.325 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:47.325 "is_configured": true, 00:14:47.325 "data_offset": 0, 00:14:47.325 "data_size": 65536 00:14:47.325 }, 00:14:47.325 { 00:14:47.325 "name": "BaseBdev2", 00:14:47.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.325 "is_configured": false, 00:14:47.325 "data_offset": 0, 00:14:47.325 "data_size": 0 00:14:47.325 }, 00:14:47.325 { 00:14:47.325 "name": "BaseBdev3", 00:14:47.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.325 "is_configured": false, 00:14:47.325 "data_offset": 0, 00:14:47.325 "data_size": 0 00:14:47.325 }, 00:14:47.325 { 00:14:47.325 "name": "BaseBdev4", 00:14:47.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.325 "is_configured": false, 00:14:47.325 "data_offset": 0, 00:14:47.325 "data_size": 0 00:14:47.325 } 00:14:47.325 ] 00:14:47.325 }' 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.325 11:03:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.585 [2024-10-29 11:03:53.039039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:47.585 BaseBdev2 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.585 [ 00:14:47.585 { 00:14:47.585 "name": "BaseBdev2", 00:14:47.585 "aliases": [ 00:14:47.585 "28c10a15-fdaf-41f4-9555-77f688856e63" 00:14:47.585 ], 00:14:47.585 "product_name": "Malloc disk", 00:14:47.585 "block_size": 512, 00:14:47.585 "num_blocks": 65536, 00:14:47.585 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:47.585 "assigned_rate_limits": { 00:14:47.585 "rw_ios_per_sec": 0, 00:14:47.585 "rw_mbytes_per_sec": 0, 00:14:47.585 "r_mbytes_per_sec": 0, 00:14:47.585 "w_mbytes_per_sec": 0 00:14:47.585 }, 00:14:47.585 "claimed": true, 00:14:47.585 "claim_type": "exclusive_write", 00:14:47.585 "zoned": false, 00:14:47.585 "supported_io_types": { 00:14:47.585 "read": true, 00:14:47.585 "write": true, 00:14:47.585 "unmap": true, 00:14:47.585 "flush": true, 00:14:47.585 "reset": true, 00:14:47.585 "nvme_admin": false, 00:14:47.585 "nvme_io": false, 00:14:47.585 "nvme_io_md": false, 00:14:47.585 "write_zeroes": true, 00:14:47.585 "zcopy": true, 00:14:47.585 "get_zone_info": false, 00:14:47.585 "zone_management": false, 00:14:47.585 "zone_append": false, 00:14:47.585 "compare": false, 00:14:47.585 "compare_and_write": false, 00:14:47.585 "abort": true, 00:14:47.585 "seek_hole": false, 00:14:47.585 "seek_data": false, 00:14:47.585 "copy": true, 00:14:47.585 "nvme_iov_md": false 00:14:47.585 }, 00:14:47.585 "memory_domains": [ 00:14:47.585 { 00:14:47.585 "dma_device_id": "system", 00:14:47.585 "dma_device_type": 1 00:14:47.585 }, 00:14:47.585 { 00:14:47.585 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:47.585 "dma_device_type": 2 00:14:47.585 } 00:14:47.585 ], 00:14:47.585 "driver_specific": {} 00:14:47.585 } 00:14:47.585 ] 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:47.585 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:47.586 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.845 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.845 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.845 "name": "Existed_Raid", 00:14:47.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.845 "strip_size_kb": 64, 00:14:47.845 "state": "configuring", 00:14:47.845 "raid_level": "raid5f", 00:14:47.845 "superblock": false, 00:14:47.845 "num_base_bdevs": 4, 00:14:47.845 "num_base_bdevs_discovered": 2, 00:14:47.845 "num_base_bdevs_operational": 4, 00:14:47.845 "base_bdevs_list": [ 00:14:47.845 { 00:14:47.845 "name": "BaseBdev1", 00:14:47.845 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:47.845 "is_configured": true, 00:14:47.845 "data_offset": 0, 00:14:47.845 "data_size": 65536 00:14:47.845 }, 00:14:47.845 { 00:14:47.845 "name": "BaseBdev2", 00:14:47.845 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:47.845 "is_configured": true, 00:14:47.845 "data_offset": 0, 00:14:47.845 "data_size": 65536 00:14:47.845 }, 00:14:47.845 { 00:14:47.845 "name": "BaseBdev3", 00:14:47.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.845 "is_configured": false, 00:14:47.845 "data_offset": 0, 00:14:47.845 "data_size": 0 00:14:47.845 }, 00:14:47.845 { 00:14:47.845 "name": "BaseBdev4", 00:14:47.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.845 "is_configured": false, 00:14:47.845 "data_offset": 0, 00:14:47.845 "data_size": 0 00:14:47.845 } 00:14:47.845 ] 00:14:47.845 }' 00:14:47.845 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.845 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 [2024-10-29 11:03:53.521192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:48.104 BaseBdev3 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 [ 00:14:48.104 { 00:14:48.104 "name": "BaseBdev3", 00:14:48.104 "aliases": [ 00:14:48.104 "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498" 00:14:48.104 ], 00:14:48.104 "product_name": "Malloc disk", 00:14:48.104 "block_size": 512, 00:14:48.104 "num_blocks": 65536, 00:14:48.104 "uuid": "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498", 00:14:48.104 "assigned_rate_limits": { 00:14:48.104 "rw_ios_per_sec": 0, 00:14:48.104 "rw_mbytes_per_sec": 0, 00:14:48.104 "r_mbytes_per_sec": 0, 00:14:48.104 "w_mbytes_per_sec": 0 00:14:48.104 }, 00:14:48.104 "claimed": true, 00:14:48.104 "claim_type": "exclusive_write", 00:14:48.104 "zoned": false, 00:14:48.104 "supported_io_types": { 00:14:48.104 "read": true, 00:14:48.104 "write": true, 00:14:48.104 "unmap": true, 00:14:48.104 "flush": true, 00:14:48.104 "reset": true, 00:14:48.104 "nvme_admin": false, 00:14:48.104 "nvme_io": false, 00:14:48.104 "nvme_io_md": false, 00:14:48.104 "write_zeroes": true, 00:14:48.104 "zcopy": true, 00:14:48.104 "get_zone_info": false, 00:14:48.104 "zone_management": false, 00:14:48.104 "zone_append": false, 00:14:48.104 "compare": false, 00:14:48.104 "compare_and_write": false, 00:14:48.104 "abort": true, 00:14:48.104 "seek_hole": false, 00:14:48.104 "seek_data": false, 00:14:48.104 "copy": true, 00:14:48.104 "nvme_iov_md": false 00:14:48.104 }, 00:14:48.104 "memory_domains": [ 00:14:48.104 { 00:14:48.104 "dma_device_id": "system", 00:14:48.104 "dma_device_type": 1 00:14:48.104 }, 00:14:48.104 { 00:14:48.104 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.104 "dma_device_type": 2 00:14:48.104 } 00:14:48.104 ], 00:14:48.104 "driver_specific": {} 00:14:48.104 } 00:14:48.104 ] 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.104 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.364 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.364 "name": "Existed_Raid", 00:14:48.364 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.364 "strip_size_kb": 64, 00:14:48.364 "state": "configuring", 00:14:48.364 "raid_level": "raid5f", 00:14:48.364 "superblock": false, 00:14:48.364 "num_base_bdevs": 4, 00:14:48.364 "num_base_bdevs_discovered": 3, 00:14:48.364 "num_base_bdevs_operational": 4, 00:14:48.364 "base_bdevs_list": [ 00:14:48.364 { 00:14:48.364 "name": "BaseBdev1", 00:14:48.364 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:48.364 "is_configured": true, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 65536 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": "BaseBdev2", 00:14:48.364 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:48.364 "is_configured": true, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 65536 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": "BaseBdev3", 00:14:48.364 "uuid": "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498", 00:14:48.364 "is_configured": true, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 65536 00:14:48.364 }, 00:14:48.364 { 00:14:48.364 "name": "BaseBdev4", 00:14:48.364 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.364 "is_configured": false, 00:14:48.364 "data_offset": 0, 00:14:48.364 "data_size": 0 00:14:48.364 } 00:14:48.364 ] 00:14:48.364 }' 00:14:48.364 11:03:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.364 11:03:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 [2024-10-29 11:03:54.031523] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:48.624 [2024-10-29 11:03:54.031588] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:48.624 [2024-10-29 11:03:54.031598] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:48.624 [2024-10-29 11:03:54.031941] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:48.624 [2024-10-29 11:03:54.032456] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:48.624 [2024-10-29 11:03:54.032485] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:14:48.624 [2024-10-29 11:03:54.032722] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:48.624 BaseBdev4 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.624 [ 00:14:48.624 { 00:14:48.624 "name": "BaseBdev4", 00:14:48.624 "aliases": [ 00:14:48.624 "b0029e75-508a-4d1e-9a8b-149a476d85fa" 00:14:48.624 ], 00:14:48.624 "product_name": "Malloc disk", 00:14:48.624 "block_size": 512, 00:14:48.624 "num_blocks": 65536, 00:14:48.624 "uuid": "b0029e75-508a-4d1e-9a8b-149a476d85fa", 00:14:48.624 "assigned_rate_limits": { 00:14:48.624 "rw_ios_per_sec": 0, 00:14:48.624 "rw_mbytes_per_sec": 0, 00:14:48.624 "r_mbytes_per_sec": 0, 00:14:48.624 "w_mbytes_per_sec": 0 00:14:48.624 }, 00:14:48.624 "claimed": true, 00:14:48.624 "claim_type": "exclusive_write", 00:14:48.624 "zoned": false, 00:14:48.624 "supported_io_types": { 00:14:48.624 "read": true, 00:14:48.624 "write": true, 00:14:48.624 "unmap": true, 00:14:48.624 "flush": true, 00:14:48.624 "reset": true, 00:14:48.624 "nvme_admin": false, 00:14:48.624 "nvme_io": false, 00:14:48.624 "nvme_io_md": false, 00:14:48.624 "write_zeroes": true, 00:14:48.624 "zcopy": true, 00:14:48.624 "get_zone_info": false, 00:14:48.624 "zone_management": false, 00:14:48.624 "zone_append": false, 00:14:48.624 "compare": false, 00:14:48.624 "compare_and_write": false, 00:14:48.624 "abort": true, 00:14:48.624 "seek_hole": false, 00:14:48.624 "seek_data": false, 00:14:48.624 "copy": true, 00:14:48.624 "nvme_iov_md": false 00:14:48.624 }, 00:14:48.624 "memory_domains": [ 00:14:48.624 { 00:14:48.624 "dma_device_id": "system", 00:14:48.624 "dma_device_type": 1 00:14:48.624 }, 00:14:48.624 { 00:14:48.624 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:48.624 "dma_device_type": 2 00:14:48.624 } 00:14:48.624 ], 00:14:48.624 "driver_specific": {} 00:14:48.624 } 00:14:48.624 ] 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:48.624 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.625 "name": "Existed_Raid", 00:14:48.625 "uuid": "8d3dcd7d-9a6f-4fc3-a567-2ef897ff3e0b", 00:14:48.625 "strip_size_kb": 64, 00:14:48.625 "state": "online", 00:14:48.625 "raid_level": "raid5f", 00:14:48.625 "superblock": false, 00:14:48.625 "num_base_bdevs": 4, 00:14:48.625 "num_base_bdevs_discovered": 4, 00:14:48.625 "num_base_bdevs_operational": 4, 00:14:48.625 "base_bdevs_list": [ 00:14:48.625 { 00:14:48.625 "name": "BaseBdev1", 00:14:48.625 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:48.625 "is_configured": true, 00:14:48.625 "data_offset": 0, 00:14:48.625 "data_size": 65536 00:14:48.625 }, 00:14:48.625 { 00:14:48.625 "name": "BaseBdev2", 00:14:48.625 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:48.625 "is_configured": true, 00:14:48.625 "data_offset": 0, 00:14:48.625 "data_size": 65536 00:14:48.625 }, 00:14:48.625 { 00:14:48.625 "name": "BaseBdev3", 00:14:48.625 "uuid": "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498", 00:14:48.625 "is_configured": true, 00:14:48.625 "data_offset": 0, 00:14:48.625 "data_size": 65536 00:14:48.625 }, 00:14:48.625 { 00:14:48.625 "name": "BaseBdev4", 00:14:48.625 "uuid": "b0029e75-508a-4d1e-9a8b-149a476d85fa", 00:14:48.625 "is_configured": true, 00:14:48.625 "data_offset": 0, 00:14:48.625 "data_size": 65536 00:14:48.625 } 00:14:48.625 ] 00:14:48.625 }' 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.625 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.194 [2024-10-29 11:03:54.534928] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:49.194 "name": "Existed_Raid", 00:14:49.194 "aliases": [ 00:14:49.194 "8d3dcd7d-9a6f-4fc3-a567-2ef897ff3e0b" 00:14:49.194 ], 00:14:49.194 "product_name": "Raid Volume", 00:14:49.194 "block_size": 512, 00:14:49.194 "num_blocks": 196608, 00:14:49.194 "uuid": "8d3dcd7d-9a6f-4fc3-a567-2ef897ff3e0b", 00:14:49.194 "assigned_rate_limits": { 00:14:49.194 "rw_ios_per_sec": 0, 00:14:49.194 "rw_mbytes_per_sec": 0, 00:14:49.194 "r_mbytes_per_sec": 0, 00:14:49.194 "w_mbytes_per_sec": 0 00:14:49.194 }, 00:14:49.194 "claimed": false, 00:14:49.194 "zoned": false, 00:14:49.194 "supported_io_types": { 00:14:49.194 "read": true, 00:14:49.194 "write": true, 00:14:49.194 "unmap": false, 00:14:49.194 "flush": false, 00:14:49.194 "reset": true, 00:14:49.194 "nvme_admin": false, 00:14:49.194 "nvme_io": false, 00:14:49.194 "nvme_io_md": false, 00:14:49.194 "write_zeroes": true, 00:14:49.194 "zcopy": false, 00:14:49.194 "get_zone_info": false, 00:14:49.194 "zone_management": false, 00:14:49.194 "zone_append": false, 00:14:49.194 "compare": false, 00:14:49.194 "compare_and_write": false, 00:14:49.194 "abort": false, 00:14:49.194 "seek_hole": false, 00:14:49.194 "seek_data": false, 00:14:49.194 "copy": false, 00:14:49.194 "nvme_iov_md": false 00:14:49.194 }, 00:14:49.194 "driver_specific": { 00:14:49.194 "raid": { 00:14:49.194 "uuid": "8d3dcd7d-9a6f-4fc3-a567-2ef897ff3e0b", 00:14:49.194 "strip_size_kb": 64, 00:14:49.194 "state": "online", 00:14:49.194 "raid_level": "raid5f", 00:14:49.194 "superblock": false, 00:14:49.194 "num_base_bdevs": 4, 00:14:49.194 "num_base_bdevs_discovered": 4, 00:14:49.194 "num_base_bdevs_operational": 4, 00:14:49.194 "base_bdevs_list": [ 00:14:49.194 { 00:14:49.194 "name": "BaseBdev1", 00:14:49.194 "uuid": "ba01531a-7532-4fb6-bd16-ff0890af2ee2", 00:14:49.194 "is_configured": true, 00:14:49.194 "data_offset": 0, 00:14:49.194 "data_size": 65536 00:14:49.194 }, 00:14:49.194 { 00:14:49.194 "name": "BaseBdev2", 00:14:49.194 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:49.194 "is_configured": true, 00:14:49.194 "data_offset": 0, 00:14:49.194 "data_size": 65536 00:14:49.194 }, 00:14:49.194 { 00:14:49.194 "name": "BaseBdev3", 00:14:49.194 "uuid": "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498", 00:14:49.194 "is_configured": true, 00:14:49.194 "data_offset": 0, 00:14:49.194 "data_size": 65536 00:14:49.194 }, 00:14:49.194 { 00:14:49.194 "name": "BaseBdev4", 00:14:49.194 "uuid": "b0029e75-508a-4d1e-9a8b-149a476d85fa", 00:14:49.194 "is_configured": true, 00:14:49.194 "data_offset": 0, 00:14:49.194 "data_size": 65536 00:14:49.194 } 00:14:49.194 ] 00:14:49.194 } 00:14:49.194 } 00:14:49.194 }' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:49.194 BaseBdev2 00:14:49.194 BaseBdev3 00:14:49.194 BaseBdev4' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.194 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:49.195 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.195 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.195 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.195 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.454 [2024-10-29 11:03:54.846272] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.454 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.454 "name": "Existed_Raid", 00:14:49.454 "uuid": "8d3dcd7d-9a6f-4fc3-a567-2ef897ff3e0b", 00:14:49.454 "strip_size_kb": 64, 00:14:49.454 "state": "online", 00:14:49.454 "raid_level": "raid5f", 00:14:49.454 "superblock": false, 00:14:49.454 "num_base_bdevs": 4, 00:14:49.454 "num_base_bdevs_discovered": 3, 00:14:49.454 "num_base_bdevs_operational": 3, 00:14:49.454 "base_bdevs_list": [ 00:14:49.454 { 00:14:49.454 "name": null, 00:14:49.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.455 "is_configured": false, 00:14:49.455 "data_offset": 0, 00:14:49.455 "data_size": 65536 00:14:49.455 }, 00:14:49.455 { 00:14:49.455 "name": "BaseBdev2", 00:14:49.455 "uuid": "28c10a15-fdaf-41f4-9555-77f688856e63", 00:14:49.455 "is_configured": true, 00:14:49.455 "data_offset": 0, 00:14:49.455 "data_size": 65536 00:14:49.455 }, 00:14:49.455 { 00:14:49.455 "name": "BaseBdev3", 00:14:49.455 "uuid": "7bdc9574-83f9-4d2d-ad57-b7bad2fd0498", 00:14:49.455 "is_configured": true, 00:14:49.455 "data_offset": 0, 00:14:49.455 "data_size": 65536 00:14:49.455 }, 00:14:49.455 { 00:14:49.455 "name": "BaseBdev4", 00:14:49.455 "uuid": "b0029e75-508a-4d1e-9a8b-149a476d85fa", 00:14:49.455 "is_configured": true, 00:14:49.455 "data_offset": 0, 00:14:49.455 "data_size": 65536 00:14:49.455 } 00:14:49.455 ] 00:14:49.455 }' 00:14:49.455 11:03:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.455 11:03:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 [2024-10-29 11:03:55.404541] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:50.025 [2024-10-29 11:03:55.404642] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:50.025 [2024-10-29 11:03:55.415924] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 [2024-10-29 11:03:55.459869] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.025 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.025 [2024-10-29 11:03:55.515186] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:50.025 [2024-10-29 11:03:55.515255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 BaseBdev2 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 [ 00:14:50.286 { 00:14:50.286 "name": "BaseBdev2", 00:14:50.286 "aliases": [ 00:14:50.286 "d14673de-3c73-4731-b568-bfd54eed1522" 00:14:50.286 ], 00:14:50.286 "product_name": "Malloc disk", 00:14:50.286 "block_size": 512, 00:14:50.286 "num_blocks": 65536, 00:14:50.286 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:50.286 "assigned_rate_limits": { 00:14:50.286 "rw_ios_per_sec": 0, 00:14:50.286 "rw_mbytes_per_sec": 0, 00:14:50.286 "r_mbytes_per_sec": 0, 00:14:50.286 "w_mbytes_per_sec": 0 00:14:50.286 }, 00:14:50.286 "claimed": false, 00:14:50.286 "zoned": false, 00:14:50.286 "supported_io_types": { 00:14:50.286 "read": true, 00:14:50.286 "write": true, 00:14:50.286 "unmap": true, 00:14:50.286 "flush": true, 00:14:50.286 "reset": true, 00:14:50.286 "nvme_admin": false, 00:14:50.286 "nvme_io": false, 00:14:50.286 "nvme_io_md": false, 00:14:50.286 "write_zeroes": true, 00:14:50.286 "zcopy": true, 00:14:50.286 "get_zone_info": false, 00:14:50.286 "zone_management": false, 00:14:50.286 "zone_append": false, 00:14:50.286 "compare": false, 00:14:50.286 "compare_and_write": false, 00:14:50.286 "abort": true, 00:14:50.286 "seek_hole": false, 00:14:50.286 "seek_data": false, 00:14:50.286 "copy": true, 00:14:50.286 "nvme_iov_md": false 00:14:50.286 }, 00:14:50.286 "memory_domains": [ 00:14:50.286 { 00:14:50.286 "dma_device_id": "system", 00:14:50.286 "dma_device_type": 1 00:14:50.286 }, 00:14:50.286 { 00:14:50.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.286 "dma_device_type": 2 00:14:50.286 } 00:14:50.286 ], 00:14:50.286 "driver_specific": {} 00:14:50.286 } 00:14:50.286 ] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.286 BaseBdev3 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:50.286 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 [ 00:14:50.287 { 00:14:50.287 "name": "BaseBdev3", 00:14:50.287 "aliases": [ 00:14:50.287 "6752a6b9-d205-4486-b10f-485faee1d07e" 00:14:50.287 ], 00:14:50.287 "product_name": "Malloc disk", 00:14:50.287 "block_size": 512, 00:14:50.287 "num_blocks": 65536, 00:14:50.287 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:50.287 "assigned_rate_limits": { 00:14:50.287 "rw_ios_per_sec": 0, 00:14:50.287 "rw_mbytes_per_sec": 0, 00:14:50.287 "r_mbytes_per_sec": 0, 00:14:50.287 "w_mbytes_per_sec": 0 00:14:50.287 }, 00:14:50.287 "claimed": false, 00:14:50.287 "zoned": false, 00:14:50.287 "supported_io_types": { 00:14:50.287 "read": true, 00:14:50.287 "write": true, 00:14:50.287 "unmap": true, 00:14:50.287 "flush": true, 00:14:50.287 "reset": true, 00:14:50.287 "nvme_admin": false, 00:14:50.287 "nvme_io": false, 00:14:50.287 "nvme_io_md": false, 00:14:50.287 "write_zeroes": true, 00:14:50.287 "zcopy": true, 00:14:50.287 "get_zone_info": false, 00:14:50.287 "zone_management": false, 00:14:50.287 "zone_append": false, 00:14:50.287 "compare": false, 00:14:50.287 "compare_and_write": false, 00:14:50.287 "abort": true, 00:14:50.287 "seek_hole": false, 00:14:50.287 "seek_data": false, 00:14:50.287 "copy": true, 00:14:50.287 "nvme_iov_md": false 00:14:50.287 }, 00:14:50.287 "memory_domains": [ 00:14:50.287 { 00:14:50.287 "dma_device_id": "system", 00:14:50.287 "dma_device_type": 1 00:14:50.287 }, 00:14:50.287 { 00:14:50.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.287 "dma_device_type": 2 00:14:50.287 } 00:14:50.287 ], 00:14:50.287 "driver_specific": {} 00:14:50.287 } 00:14:50.287 ] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 BaseBdev4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 [ 00:14:50.287 { 00:14:50.287 "name": "BaseBdev4", 00:14:50.287 "aliases": [ 00:14:50.287 "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf" 00:14:50.287 ], 00:14:50.287 "product_name": "Malloc disk", 00:14:50.287 "block_size": 512, 00:14:50.287 "num_blocks": 65536, 00:14:50.287 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:50.287 "assigned_rate_limits": { 00:14:50.287 "rw_ios_per_sec": 0, 00:14:50.287 "rw_mbytes_per_sec": 0, 00:14:50.287 "r_mbytes_per_sec": 0, 00:14:50.287 "w_mbytes_per_sec": 0 00:14:50.287 }, 00:14:50.287 "claimed": false, 00:14:50.287 "zoned": false, 00:14:50.287 "supported_io_types": { 00:14:50.287 "read": true, 00:14:50.287 "write": true, 00:14:50.287 "unmap": true, 00:14:50.287 "flush": true, 00:14:50.287 "reset": true, 00:14:50.287 "nvme_admin": false, 00:14:50.287 "nvme_io": false, 00:14:50.287 "nvme_io_md": false, 00:14:50.287 "write_zeroes": true, 00:14:50.287 "zcopy": true, 00:14:50.287 "get_zone_info": false, 00:14:50.287 "zone_management": false, 00:14:50.287 "zone_append": false, 00:14:50.287 "compare": false, 00:14:50.287 "compare_and_write": false, 00:14:50.287 "abort": true, 00:14:50.287 "seek_hole": false, 00:14:50.287 "seek_data": false, 00:14:50.287 "copy": true, 00:14:50.287 "nvme_iov_md": false 00:14:50.287 }, 00:14:50.287 "memory_domains": [ 00:14:50.287 { 00:14:50.287 "dma_device_id": "system", 00:14:50.287 "dma_device_type": 1 00:14:50.287 }, 00:14:50.287 { 00:14:50.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.287 "dma_device_type": 2 00:14:50.287 } 00:14:50.287 ], 00:14:50.287 "driver_specific": {} 00:14:50.287 } 00:14:50.287 ] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 [2024-10-29 11:03:55.730756] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:50.287 [2024-10-29 11:03:55.730803] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:50.287 [2024-10-29 11:03:55.730842] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:50.287 [2024-10-29 11:03:55.732716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:50.287 [2024-10-29 11:03:55.732774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.287 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.548 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.548 "name": "Existed_Raid", 00:14:50.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.548 "strip_size_kb": 64, 00:14:50.548 "state": "configuring", 00:14:50.548 "raid_level": "raid5f", 00:14:50.548 "superblock": false, 00:14:50.548 "num_base_bdevs": 4, 00:14:50.548 "num_base_bdevs_discovered": 3, 00:14:50.548 "num_base_bdevs_operational": 4, 00:14:50.548 "base_bdevs_list": [ 00:14:50.548 { 00:14:50.548 "name": "BaseBdev1", 00:14:50.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.548 "is_configured": false, 00:14:50.548 "data_offset": 0, 00:14:50.548 "data_size": 0 00:14:50.548 }, 00:14:50.548 { 00:14:50.548 "name": "BaseBdev2", 00:14:50.548 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:50.548 "is_configured": true, 00:14:50.548 "data_offset": 0, 00:14:50.548 "data_size": 65536 00:14:50.548 }, 00:14:50.548 { 00:14:50.548 "name": "BaseBdev3", 00:14:50.548 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:50.548 "is_configured": true, 00:14:50.548 "data_offset": 0, 00:14:50.548 "data_size": 65536 00:14:50.548 }, 00:14:50.548 { 00:14:50.548 "name": "BaseBdev4", 00:14:50.548 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:50.548 "is_configured": true, 00:14:50.548 "data_offset": 0, 00:14:50.548 "data_size": 65536 00:14:50.548 } 00:14:50.548 ] 00:14:50.548 }' 00:14:50.548 11:03:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.548 11:03:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.808 [2024-10-29 11:03:56.169961] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.808 "name": "Existed_Raid", 00:14:50.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.808 "strip_size_kb": 64, 00:14:50.808 "state": "configuring", 00:14:50.808 "raid_level": "raid5f", 00:14:50.808 "superblock": false, 00:14:50.808 "num_base_bdevs": 4, 00:14:50.808 "num_base_bdevs_discovered": 2, 00:14:50.808 "num_base_bdevs_operational": 4, 00:14:50.808 "base_bdevs_list": [ 00:14:50.808 { 00:14:50.808 "name": "BaseBdev1", 00:14:50.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.808 "is_configured": false, 00:14:50.808 "data_offset": 0, 00:14:50.808 "data_size": 0 00:14:50.808 }, 00:14:50.808 { 00:14:50.808 "name": null, 00:14:50.808 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:50.808 "is_configured": false, 00:14:50.808 "data_offset": 0, 00:14:50.808 "data_size": 65536 00:14:50.808 }, 00:14:50.808 { 00:14:50.808 "name": "BaseBdev3", 00:14:50.808 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:50.808 "is_configured": true, 00:14:50.808 "data_offset": 0, 00:14:50.808 "data_size": 65536 00:14:50.808 }, 00:14:50.808 { 00:14:50.808 "name": "BaseBdev4", 00:14:50.808 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:50.808 "is_configured": true, 00:14:50.808 "data_offset": 0, 00:14:50.808 "data_size": 65536 00:14:50.808 } 00:14:50.808 ] 00:14:50.808 }' 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.808 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 [2024-10-29 11:03:56.672481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:51.379 BaseBdev1 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 [ 00:14:51.379 { 00:14:51.379 "name": "BaseBdev1", 00:14:51.379 "aliases": [ 00:14:51.379 "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9" 00:14:51.379 ], 00:14:51.379 "product_name": "Malloc disk", 00:14:51.379 "block_size": 512, 00:14:51.379 "num_blocks": 65536, 00:14:51.379 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:51.379 "assigned_rate_limits": { 00:14:51.379 "rw_ios_per_sec": 0, 00:14:51.379 "rw_mbytes_per_sec": 0, 00:14:51.379 "r_mbytes_per_sec": 0, 00:14:51.379 "w_mbytes_per_sec": 0 00:14:51.379 }, 00:14:51.379 "claimed": true, 00:14:51.379 "claim_type": "exclusive_write", 00:14:51.379 "zoned": false, 00:14:51.379 "supported_io_types": { 00:14:51.379 "read": true, 00:14:51.379 "write": true, 00:14:51.379 "unmap": true, 00:14:51.379 "flush": true, 00:14:51.379 "reset": true, 00:14:51.379 "nvme_admin": false, 00:14:51.379 "nvme_io": false, 00:14:51.379 "nvme_io_md": false, 00:14:51.379 "write_zeroes": true, 00:14:51.379 "zcopy": true, 00:14:51.379 "get_zone_info": false, 00:14:51.379 "zone_management": false, 00:14:51.379 "zone_append": false, 00:14:51.379 "compare": false, 00:14:51.379 "compare_and_write": false, 00:14:51.379 "abort": true, 00:14:51.379 "seek_hole": false, 00:14:51.379 "seek_data": false, 00:14:51.379 "copy": true, 00:14:51.379 "nvme_iov_md": false 00:14:51.379 }, 00:14:51.379 "memory_domains": [ 00:14:51.379 { 00:14:51.379 "dma_device_id": "system", 00:14:51.379 "dma_device_type": 1 00:14:51.379 }, 00:14:51.379 { 00:14:51.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:51.379 "dma_device_type": 2 00:14:51.379 } 00:14:51.379 ], 00:14:51.379 "driver_specific": {} 00:14:51.379 } 00:14:51.379 ] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.379 "name": "Existed_Raid", 00:14:51.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.379 "strip_size_kb": 64, 00:14:51.379 "state": "configuring", 00:14:51.379 "raid_level": "raid5f", 00:14:51.379 "superblock": false, 00:14:51.379 "num_base_bdevs": 4, 00:14:51.379 "num_base_bdevs_discovered": 3, 00:14:51.379 "num_base_bdevs_operational": 4, 00:14:51.379 "base_bdevs_list": [ 00:14:51.379 { 00:14:51.379 "name": "BaseBdev1", 00:14:51.379 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:51.379 "is_configured": true, 00:14:51.379 "data_offset": 0, 00:14:51.379 "data_size": 65536 00:14:51.379 }, 00:14:51.379 { 00:14:51.379 "name": null, 00:14:51.379 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:51.379 "is_configured": false, 00:14:51.379 "data_offset": 0, 00:14:51.379 "data_size": 65536 00:14:51.379 }, 00:14:51.379 { 00:14:51.379 "name": "BaseBdev3", 00:14:51.379 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:51.379 "is_configured": true, 00:14:51.379 "data_offset": 0, 00:14:51.379 "data_size": 65536 00:14:51.379 }, 00:14:51.379 { 00:14:51.379 "name": "BaseBdev4", 00:14:51.379 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:51.379 "is_configured": true, 00:14:51.379 "data_offset": 0, 00:14:51.379 "data_size": 65536 00:14:51.379 } 00:14:51.379 ] 00:14:51.379 }' 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.379 11:03:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.950 [2024-10-29 11:03:57.212006] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.950 "name": "Existed_Raid", 00:14:51.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.950 "strip_size_kb": 64, 00:14:51.950 "state": "configuring", 00:14:51.950 "raid_level": "raid5f", 00:14:51.950 "superblock": false, 00:14:51.950 "num_base_bdevs": 4, 00:14:51.950 "num_base_bdevs_discovered": 2, 00:14:51.950 "num_base_bdevs_operational": 4, 00:14:51.950 "base_bdevs_list": [ 00:14:51.950 { 00:14:51.950 "name": "BaseBdev1", 00:14:51.950 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:51.950 "is_configured": true, 00:14:51.950 "data_offset": 0, 00:14:51.950 "data_size": 65536 00:14:51.950 }, 00:14:51.950 { 00:14:51.950 "name": null, 00:14:51.950 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:51.950 "is_configured": false, 00:14:51.950 "data_offset": 0, 00:14:51.950 "data_size": 65536 00:14:51.950 }, 00:14:51.950 { 00:14:51.950 "name": null, 00:14:51.950 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:51.950 "is_configured": false, 00:14:51.950 "data_offset": 0, 00:14:51.950 "data_size": 65536 00:14:51.950 }, 00:14:51.950 { 00:14:51.950 "name": "BaseBdev4", 00:14:51.950 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:51.950 "is_configured": true, 00:14:51.950 "data_offset": 0, 00:14:51.950 "data_size": 65536 00:14:51.950 } 00:14:51.950 ] 00:14:51.950 }' 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.950 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.212 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.212 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:52.212 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.212 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.213 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.476 [2024-10-29 11:03:57.731531] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.476 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.476 "name": "Existed_Raid", 00:14:52.476 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.477 "strip_size_kb": 64, 00:14:52.477 "state": "configuring", 00:14:52.477 "raid_level": "raid5f", 00:14:52.477 "superblock": false, 00:14:52.477 "num_base_bdevs": 4, 00:14:52.477 "num_base_bdevs_discovered": 3, 00:14:52.477 "num_base_bdevs_operational": 4, 00:14:52.477 "base_bdevs_list": [ 00:14:52.477 { 00:14:52.477 "name": "BaseBdev1", 00:14:52.477 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:52.477 "is_configured": true, 00:14:52.477 "data_offset": 0, 00:14:52.477 "data_size": 65536 00:14:52.477 }, 00:14:52.477 { 00:14:52.477 "name": null, 00:14:52.477 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:52.477 "is_configured": false, 00:14:52.477 "data_offset": 0, 00:14:52.477 "data_size": 65536 00:14:52.477 }, 00:14:52.477 { 00:14:52.477 "name": "BaseBdev3", 00:14:52.477 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:52.477 "is_configured": true, 00:14:52.477 "data_offset": 0, 00:14:52.477 "data_size": 65536 00:14:52.477 }, 00:14:52.477 { 00:14:52.477 "name": "BaseBdev4", 00:14:52.477 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:52.477 "is_configured": true, 00:14:52.477 "data_offset": 0, 00:14:52.477 "data_size": 65536 00:14:52.477 } 00:14:52.477 ] 00:14:52.477 }' 00:14:52.477 11:03:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.477 11:03:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.744 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.744 [2024-10-29 11:03:58.234641] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.015 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.015 "name": "Existed_Raid", 00:14:53.015 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.015 "strip_size_kb": 64, 00:14:53.015 "state": "configuring", 00:14:53.015 "raid_level": "raid5f", 00:14:53.015 "superblock": false, 00:14:53.015 "num_base_bdevs": 4, 00:14:53.015 "num_base_bdevs_discovered": 2, 00:14:53.015 "num_base_bdevs_operational": 4, 00:14:53.015 "base_bdevs_list": [ 00:14:53.015 { 00:14:53.015 "name": null, 00:14:53.015 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:53.015 "is_configured": false, 00:14:53.015 "data_offset": 0, 00:14:53.015 "data_size": 65536 00:14:53.015 }, 00:14:53.015 { 00:14:53.015 "name": null, 00:14:53.015 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:53.015 "is_configured": false, 00:14:53.015 "data_offset": 0, 00:14:53.015 "data_size": 65536 00:14:53.015 }, 00:14:53.015 { 00:14:53.015 "name": "BaseBdev3", 00:14:53.015 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:53.015 "is_configured": true, 00:14:53.015 "data_offset": 0, 00:14:53.015 "data_size": 65536 00:14:53.015 }, 00:14:53.015 { 00:14:53.015 "name": "BaseBdev4", 00:14:53.015 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:53.015 "is_configured": true, 00:14:53.015 "data_offset": 0, 00:14:53.015 "data_size": 65536 00:14:53.016 } 00:14:53.016 ] 00:14:53.016 }' 00:14:53.016 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.016 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.282 [2024-10-29 11:03:58.716496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.282 "name": "Existed_Raid", 00:14:53.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.282 "strip_size_kb": 64, 00:14:53.282 "state": "configuring", 00:14:53.282 "raid_level": "raid5f", 00:14:53.282 "superblock": false, 00:14:53.282 "num_base_bdevs": 4, 00:14:53.282 "num_base_bdevs_discovered": 3, 00:14:53.282 "num_base_bdevs_operational": 4, 00:14:53.282 "base_bdevs_list": [ 00:14:53.282 { 00:14:53.282 "name": null, 00:14:53.282 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:53.282 "is_configured": false, 00:14:53.282 "data_offset": 0, 00:14:53.282 "data_size": 65536 00:14:53.282 }, 00:14:53.282 { 00:14:53.282 "name": "BaseBdev2", 00:14:53.282 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:53.282 "is_configured": true, 00:14:53.282 "data_offset": 0, 00:14:53.282 "data_size": 65536 00:14:53.282 }, 00:14:53.282 { 00:14:53.282 "name": "BaseBdev3", 00:14:53.282 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:53.282 "is_configured": true, 00:14:53.282 "data_offset": 0, 00:14:53.282 "data_size": 65536 00:14:53.282 }, 00:14:53.282 { 00:14:53.282 "name": "BaseBdev4", 00:14:53.282 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:53.282 "is_configured": true, 00:14:53.282 "data_offset": 0, 00:14:53.282 "data_size": 65536 00:14:53.282 } 00:14:53.282 ] 00:14:53.282 }' 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.282 11:03:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.851 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 [2024-10-29 11:03:59.234872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:53.852 [2024-10-29 11:03:59.234947] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:14:53.852 [2024-10-29 11:03:59.234957] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:53.852 [2024-10-29 11:03:59.235230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:14:53.852 [2024-10-29 11:03:59.235741] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:14:53.852 [2024-10-29 11:03:59.235769] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:14:53.852 [2024-10-29 11:03:59.235979] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:53.852 NewBaseBdev 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local i 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 [ 00:14:53.852 { 00:14:53.852 "name": "NewBaseBdev", 00:14:53.852 "aliases": [ 00:14:53.852 "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9" 00:14:53.852 ], 00:14:53.852 "product_name": "Malloc disk", 00:14:53.852 "block_size": 512, 00:14:53.852 "num_blocks": 65536, 00:14:53.852 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:53.852 "assigned_rate_limits": { 00:14:53.852 "rw_ios_per_sec": 0, 00:14:53.852 "rw_mbytes_per_sec": 0, 00:14:53.852 "r_mbytes_per_sec": 0, 00:14:53.852 "w_mbytes_per_sec": 0 00:14:53.852 }, 00:14:53.852 "claimed": true, 00:14:53.852 "claim_type": "exclusive_write", 00:14:53.852 "zoned": false, 00:14:53.852 "supported_io_types": { 00:14:53.852 "read": true, 00:14:53.852 "write": true, 00:14:53.852 "unmap": true, 00:14:53.852 "flush": true, 00:14:53.852 "reset": true, 00:14:53.852 "nvme_admin": false, 00:14:53.852 "nvme_io": false, 00:14:53.852 "nvme_io_md": false, 00:14:53.852 "write_zeroes": true, 00:14:53.852 "zcopy": true, 00:14:53.852 "get_zone_info": false, 00:14:53.852 "zone_management": false, 00:14:53.852 "zone_append": false, 00:14:53.852 "compare": false, 00:14:53.852 "compare_and_write": false, 00:14:53.852 "abort": true, 00:14:53.852 "seek_hole": false, 00:14:53.852 "seek_data": false, 00:14:53.852 "copy": true, 00:14:53.852 "nvme_iov_md": false 00:14:53.852 }, 00:14:53.852 "memory_domains": [ 00:14:53.852 { 00:14:53.852 "dma_device_id": "system", 00:14:53.852 "dma_device_type": 1 00:14:53.852 }, 00:14:53.852 { 00:14:53.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:53.852 "dma_device_type": 2 00:14:53.852 } 00:14:53.852 ], 00:14:53.852 "driver_specific": {} 00:14:53.852 } 00:14:53.852 ] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@909 -- # return 0 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.852 "name": "Existed_Raid", 00:14:53.852 "uuid": "2f0c846c-e26f-4b4d-a955-e1584709ac71", 00:14:53.852 "strip_size_kb": 64, 00:14:53.852 "state": "online", 00:14:53.852 "raid_level": "raid5f", 00:14:53.852 "superblock": false, 00:14:53.852 "num_base_bdevs": 4, 00:14:53.852 "num_base_bdevs_discovered": 4, 00:14:53.852 "num_base_bdevs_operational": 4, 00:14:53.852 "base_bdevs_list": [ 00:14:53.852 { 00:14:53.852 "name": "NewBaseBdev", 00:14:53.852 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:53.852 "is_configured": true, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.852 }, 00:14:53.852 { 00:14:53.852 "name": "BaseBdev2", 00:14:53.852 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:53.852 "is_configured": true, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.852 }, 00:14:53.852 { 00:14:53.852 "name": "BaseBdev3", 00:14:53.852 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:53.852 "is_configured": true, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.852 }, 00:14:53.852 { 00:14:53.852 "name": "BaseBdev4", 00:14:53.852 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:53.852 "is_configured": true, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.852 } 00:14:53.852 ] 00:14:53.852 }' 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.852 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:54.421 [2024-10-29 11:03:59.722338] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:54.421 "name": "Existed_Raid", 00:14:54.421 "aliases": [ 00:14:54.421 "2f0c846c-e26f-4b4d-a955-e1584709ac71" 00:14:54.421 ], 00:14:54.421 "product_name": "Raid Volume", 00:14:54.421 "block_size": 512, 00:14:54.421 "num_blocks": 196608, 00:14:54.421 "uuid": "2f0c846c-e26f-4b4d-a955-e1584709ac71", 00:14:54.421 "assigned_rate_limits": { 00:14:54.421 "rw_ios_per_sec": 0, 00:14:54.421 "rw_mbytes_per_sec": 0, 00:14:54.421 "r_mbytes_per_sec": 0, 00:14:54.421 "w_mbytes_per_sec": 0 00:14:54.421 }, 00:14:54.421 "claimed": false, 00:14:54.421 "zoned": false, 00:14:54.421 "supported_io_types": { 00:14:54.421 "read": true, 00:14:54.421 "write": true, 00:14:54.421 "unmap": false, 00:14:54.421 "flush": false, 00:14:54.421 "reset": true, 00:14:54.421 "nvme_admin": false, 00:14:54.421 "nvme_io": false, 00:14:54.421 "nvme_io_md": false, 00:14:54.421 "write_zeroes": true, 00:14:54.421 "zcopy": false, 00:14:54.421 "get_zone_info": false, 00:14:54.421 "zone_management": false, 00:14:54.421 "zone_append": false, 00:14:54.421 "compare": false, 00:14:54.421 "compare_and_write": false, 00:14:54.421 "abort": false, 00:14:54.421 "seek_hole": false, 00:14:54.421 "seek_data": false, 00:14:54.421 "copy": false, 00:14:54.421 "nvme_iov_md": false 00:14:54.421 }, 00:14:54.421 "driver_specific": { 00:14:54.421 "raid": { 00:14:54.421 "uuid": "2f0c846c-e26f-4b4d-a955-e1584709ac71", 00:14:54.421 "strip_size_kb": 64, 00:14:54.421 "state": "online", 00:14:54.421 "raid_level": "raid5f", 00:14:54.421 "superblock": false, 00:14:54.421 "num_base_bdevs": 4, 00:14:54.421 "num_base_bdevs_discovered": 4, 00:14:54.421 "num_base_bdevs_operational": 4, 00:14:54.421 "base_bdevs_list": [ 00:14:54.421 { 00:14:54.421 "name": "NewBaseBdev", 00:14:54.421 "uuid": "b6395ae9-4a06-4c8a-a62b-6ba12a62a0c9", 00:14:54.421 "is_configured": true, 00:14:54.421 "data_offset": 0, 00:14:54.421 "data_size": 65536 00:14:54.421 }, 00:14:54.421 { 00:14:54.421 "name": "BaseBdev2", 00:14:54.421 "uuid": "d14673de-3c73-4731-b568-bfd54eed1522", 00:14:54.421 "is_configured": true, 00:14:54.421 "data_offset": 0, 00:14:54.421 "data_size": 65536 00:14:54.421 }, 00:14:54.421 { 00:14:54.421 "name": "BaseBdev3", 00:14:54.421 "uuid": "6752a6b9-d205-4486-b10f-485faee1d07e", 00:14:54.421 "is_configured": true, 00:14:54.421 "data_offset": 0, 00:14:54.421 "data_size": 65536 00:14:54.421 }, 00:14:54.421 { 00:14:54.421 "name": "BaseBdev4", 00:14:54.421 "uuid": "da6fc33f-8d51-465b-8b8c-3f3dbf246cbf", 00:14:54.421 "is_configured": true, 00:14:54.421 "data_offset": 0, 00:14:54.421 "data_size": 65536 00:14:54.421 } 00:14:54.421 ] 00:14:54.421 } 00:14:54.421 } 00:14:54.421 }' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:54.421 BaseBdev2 00:14:54.421 BaseBdev3 00:14:54.421 BaseBdev4' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.421 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.680 11:03:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.680 [2024-10-29 11:04:00.037565] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:54.680 [2024-10-29 11:04:00.037641] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:54.680 [2024-10-29 11:04:00.037731] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:54.680 [2024-10-29 11:04:00.038038] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:54.680 [2024-10-29 11:04:00.038100] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 93344 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@952 -- # '[' -z 93344 ']' 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # kill -0 93344 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@957 -- # uname 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 93344 00:14:54.680 killing process with pid 93344 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 93344' 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@971 -- # kill 93344 00:14:54.680 [2024-10-29 11:04:00.085099] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:54.680 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@976 -- # wait 93344 00:14:54.680 [2024-10-29 11:04:00.125257] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:54.940 00:14:54.940 real 0m9.642s 00:14:54.940 user 0m16.430s 00:14:54.940 sys 0m2.148s 00:14:54.940 ************************************ 00:14:54.940 END TEST raid5f_state_function_test 00:14:54.940 ************************************ 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.940 11:04:00 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:54.940 11:04:00 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:14:54.940 11:04:00 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:14:54.940 11:04:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:54.940 ************************************ 00:14:54.940 START TEST raid5f_state_function_test_sb 00:14:54.940 ************************************ 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1127 -- # raid_state_function_test raid5f 4 true 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:54.940 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:54.941 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:55.201 Process raid pid: 93999 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93999 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93999' 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93999 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@833 -- # '[' -z 93999 ']' 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:55.201 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:14:55.201 11:04:00 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.201 [2024-10-29 11:04:00.542841] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:14:55.201 [2024-10-29 11:04:00.542993] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:55.461 [2024-10-29 11:04:00.719745] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:55.461 [2024-10-29 11:04:00.747031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:55.461 [2024-10-29 11:04:00.792704] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:55.461 [2024-10-29 11:04:00.792812] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@866 -- # return 0 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.031 [2024-10-29 11:04:01.371728] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:56.031 [2024-10-29 11:04:01.371790] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:56.031 [2024-10-29 11:04:01.371802] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:56.031 [2024-10-29 11:04:01.371814] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:56.031 [2024-10-29 11:04:01.371821] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:56.031 [2024-10-29 11:04:01.371834] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:56.031 [2024-10-29 11:04:01.371841] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:56.031 [2024-10-29 11:04:01.371852] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:56.031 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.032 "name": "Existed_Raid", 00:14:56.032 "uuid": "dad31f80-58d2-49c6-9342-848369571157", 00:14:56.032 "strip_size_kb": 64, 00:14:56.032 "state": "configuring", 00:14:56.032 "raid_level": "raid5f", 00:14:56.032 "superblock": true, 00:14:56.032 "num_base_bdevs": 4, 00:14:56.032 "num_base_bdevs_discovered": 0, 00:14:56.032 "num_base_bdevs_operational": 4, 00:14:56.032 "base_bdevs_list": [ 00:14:56.032 { 00:14:56.032 "name": "BaseBdev1", 00:14:56.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.032 "is_configured": false, 00:14:56.032 "data_offset": 0, 00:14:56.032 "data_size": 0 00:14:56.032 }, 00:14:56.032 { 00:14:56.032 "name": "BaseBdev2", 00:14:56.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.032 "is_configured": false, 00:14:56.032 "data_offset": 0, 00:14:56.032 "data_size": 0 00:14:56.032 }, 00:14:56.032 { 00:14:56.032 "name": "BaseBdev3", 00:14:56.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.032 "is_configured": false, 00:14:56.032 "data_offset": 0, 00:14:56.032 "data_size": 0 00:14:56.032 }, 00:14:56.032 { 00:14:56.032 "name": "BaseBdev4", 00:14:56.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.032 "is_configured": false, 00:14:56.032 "data_offset": 0, 00:14:56.032 "data_size": 0 00:14:56.032 } 00:14:56.032 ] 00:14:56.032 }' 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.032 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.297 [2024-10-29 11:04:01.750988] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:56.297 [2024-10-29 11:04:01.751095] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.297 [2024-10-29 11:04:01.762973] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:56.297 [2024-10-29 11:04:01.763067] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:56.297 [2024-10-29 11:04:01.763096] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:56.297 [2024-10-29 11:04:01.763123] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:56.297 [2024-10-29 11:04:01.763144] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:56.297 [2024-10-29 11:04:01.763184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:56.297 [2024-10-29 11:04:01.763229] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:56.297 [2024-10-29 11:04:01.763257] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.297 [2024-10-29 11:04:01.784023] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:56.297 BaseBdev1 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:56.297 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.298 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.558 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.559 [ 00:14:56.559 { 00:14:56.559 "name": "BaseBdev1", 00:14:56.559 "aliases": [ 00:14:56.559 "6080e89f-b6a9-4746-a388-0fa619605eb6" 00:14:56.559 ], 00:14:56.559 "product_name": "Malloc disk", 00:14:56.559 "block_size": 512, 00:14:56.559 "num_blocks": 65536, 00:14:56.559 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:56.559 "assigned_rate_limits": { 00:14:56.559 "rw_ios_per_sec": 0, 00:14:56.559 "rw_mbytes_per_sec": 0, 00:14:56.559 "r_mbytes_per_sec": 0, 00:14:56.559 "w_mbytes_per_sec": 0 00:14:56.559 }, 00:14:56.559 "claimed": true, 00:14:56.559 "claim_type": "exclusive_write", 00:14:56.559 "zoned": false, 00:14:56.559 "supported_io_types": { 00:14:56.559 "read": true, 00:14:56.559 "write": true, 00:14:56.559 "unmap": true, 00:14:56.559 "flush": true, 00:14:56.559 "reset": true, 00:14:56.559 "nvme_admin": false, 00:14:56.559 "nvme_io": false, 00:14:56.559 "nvme_io_md": false, 00:14:56.559 "write_zeroes": true, 00:14:56.559 "zcopy": true, 00:14:56.559 "get_zone_info": false, 00:14:56.559 "zone_management": false, 00:14:56.559 "zone_append": false, 00:14:56.559 "compare": false, 00:14:56.559 "compare_and_write": false, 00:14:56.559 "abort": true, 00:14:56.559 "seek_hole": false, 00:14:56.559 "seek_data": false, 00:14:56.559 "copy": true, 00:14:56.559 "nvme_iov_md": false 00:14:56.559 }, 00:14:56.559 "memory_domains": [ 00:14:56.559 { 00:14:56.559 "dma_device_id": "system", 00:14:56.559 "dma_device_type": 1 00:14:56.559 }, 00:14:56.559 { 00:14:56.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:56.559 "dma_device_type": 2 00:14:56.559 } 00:14:56.559 ], 00:14:56.559 "driver_specific": {} 00:14:56.559 } 00:14:56.559 ] 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.559 "name": "Existed_Raid", 00:14:56.559 "uuid": "819e68f0-c379-4d5d-aad8-116cd2398546", 00:14:56.559 "strip_size_kb": 64, 00:14:56.559 "state": "configuring", 00:14:56.559 "raid_level": "raid5f", 00:14:56.559 "superblock": true, 00:14:56.559 "num_base_bdevs": 4, 00:14:56.559 "num_base_bdevs_discovered": 1, 00:14:56.559 "num_base_bdevs_operational": 4, 00:14:56.559 "base_bdevs_list": [ 00:14:56.559 { 00:14:56.559 "name": "BaseBdev1", 00:14:56.559 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:56.559 "is_configured": true, 00:14:56.559 "data_offset": 2048, 00:14:56.559 "data_size": 63488 00:14:56.559 }, 00:14:56.559 { 00:14:56.559 "name": "BaseBdev2", 00:14:56.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.559 "is_configured": false, 00:14:56.559 "data_offset": 0, 00:14:56.559 "data_size": 0 00:14:56.559 }, 00:14:56.559 { 00:14:56.559 "name": "BaseBdev3", 00:14:56.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.559 "is_configured": false, 00:14:56.559 "data_offset": 0, 00:14:56.559 "data_size": 0 00:14:56.559 }, 00:14:56.559 { 00:14:56.559 "name": "BaseBdev4", 00:14:56.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.559 "is_configured": false, 00:14:56.559 "data_offset": 0, 00:14:56.559 "data_size": 0 00:14:56.559 } 00:14:56.559 ] 00:14:56.559 }' 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.559 11:04:01 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.818 [2024-10-29 11:04:02.199450] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:56.818 [2024-10-29 11:04:02.199495] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.818 [2024-10-29 11:04:02.211476] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:56.818 [2024-10-29 11:04:02.213399] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:56.818 [2024-10-29 11:04:02.213444] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:56.818 [2024-10-29 11:04:02.213454] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:56.818 [2024-10-29 11:04:02.213465] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:56.818 [2024-10-29 11:04:02.213474] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:56.818 [2024-10-29 11:04:02.213484] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.818 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.819 "name": "Existed_Raid", 00:14:56.819 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:56.819 "strip_size_kb": 64, 00:14:56.819 "state": "configuring", 00:14:56.819 "raid_level": "raid5f", 00:14:56.819 "superblock": true, 00:14:56.819 "num_base_bdevs": 4, 00:14:56.819 "num_base_bdevs_discovered": 1, 00:14:56.819 "num_base_bdevs_operational": 4, 00:14:56.819 "base_bdevs_list": [ 00:14:56.819 { 00:14:56.819 "name": "BaseBdev1", 00:14:56.819 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:56.819 "is_configured": true, 00:14:56.819 "data_offset": 2048, 00:14:56.819 "data_size": 63488 00:14:56.819 }, 00:14:56.819 { 00:14:56.819 "name": "BaseBdev2", 00:14:56.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.819 "is_configured": false, 00:14:56.819 "data_offset": 0, 00:14:56.819 "data_size": 0 00:14:56.819 }, 00:14:56.819 { 00:14:56.819 "name": "BaseBdev3", 00:14:56.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.819 "is_configured": false, 00:14:56.819 "data_offset": 0, 00:14:56.819 "data_size": 0 00:14:56.819 }, 00:14:56.819 { 00:14:56.819 "name": "BaseBdev4", 00:14:56.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.819 "is_configured": false, 00:14:56.819 "data_offset": 0, 00:14:56.819 "data_size": 0 00:14:56.819 } 00:14:56.819 ] 00:14:56.819 }' 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.819 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.389 [2024-10-29 11:04:02.621993] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:57.389 BaseBdev2 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:57.389 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.390 [ 00:14:57.390 { 00:14:57.390 "name": "BaseBdev2", 00:14:57.390 "aliases": [ 00:14:57.390 "23ab6273-bd28-4dee-b90c-cced9b0fe468" 00:14:57.390 ], 00:14:57.390 "product_name": "Malloc disk", 00:14:57.390 "block_size": 512, 00:14:57.390 "num_blocks": 65536, 00:14:57.390 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:57.390 "assigned_rate_limits": { 00:14:57.390 "rw_ios_per_sec": 0, 00:14:57.390 "rw_mbytes_per_sec": 0, 00:14:57.390 "r_mbytes_per_sec": 0, 00:14:57.390 "w_mbytes_per_sec": 0 00:14:57.390 }, 00:14:57.390 "claimed": true, 00:14:57.390 "claim_type": "exclusive_write", 00:14:57.390 "zoned": false, 00:14:57.390 "supported_io_types": { 00:14:57.390 "read": true, 00:14:57.390 "write": true, 00:14:57.390 "unmap": true, 00:14:57.390 "flush": true, 00:14:57.390 "reset": true, 00:14:57.390 "nvme_admin": false, 00:14:57.390 "nvme_io": false, 00:14:57.390 "nvme_io_md": false, 00:14:57.390 "write_zeroes": true, 00:14:57.390 "zcopy": true, 00:14:57.390 "get_zone_info": false, 00:14:57.390 "zone_management": false, 00:14:57.390 "zone_append": false, 00:14:57.390 "compare": false, 00:14:57.390 "compare_and_write": false, 00:14:57.390 "abort": true, 00:14:57.390 "seek_hole": false, 00:14:57.390 "seek_data": false, 00:14:57.390 "copy": true, 00:14:57.390 "nvme_iov_md": false 00:14:57.390 }, 00:14:57.390 "memory_domains": [ 00:14:57.390 { 00:14:57.390 "dma_device_id": "system", 00:14:57.390 "dma_device_type": 1 00:14:57.390 }, 00:14:57.390 { 00:14:57.390 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:57.390 "dma_device_type": 2 00:14:57.390 } 00:14:57.390 ], 00:14:57.390 "driver_specific": {} 00:14:57.390 } 00:14:57.390 ] 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.390 "name": "Existed_Raid", 00:14:57.390 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:57.390 "strip_size_kb": 64, 00:14:57.390 "state": "configuring", 00:14:57.390 "raid_level": "raid5f", 00:14:57.390 "superblock": true, 00:14:57.390 "num_base_bdevs": 4, 00:14:57.390 "num_base_bdevs_discovered": 2, 00:14:57.390 "num_base_bdevs_operational": 4, 00:14:57.390 "base_bdevs_list": [ 00:14:57.390 { 00:14:57.390 "name": "BaseBdev1", 00:14:57.390 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:57.390 "is_configured": true, 00:14:57.390 "data_offset": 2048, 00:14:57.390 "data_size": 63488 00:14:57.390 }, 00:14:57.390 { 00:14:57.390 "name": "BaseBdev2", 00:14:57.390 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:57.390 "is_configured": true, 00:14:57.390 "data_offset": 2048, 00:14:57.390 "data_size": 63488 00:14:57.390 }, 00:14:57.390 { 00:14:57.390 "name": "BaseBdev3", 00:14:57.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.390 "is_configured": false, 00:14:57.390 "data_offset": 0, 00:14:57.390 "data_size": 0 00:14:57.390 }, 00:14:57.390 { 00:14:57.390 "name": "BaseBdev4", 00:14:57.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.390 "is_configured": false, 00:14:57.390 "data_offset": 0, 00:14:57.390 "data_size": 0 00:14:57.390 } 00:14:57.390 ] 00:14:57.390 }' 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.390 11:04:02 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.650 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:57.650 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.650 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.911 [2024-10-29 11:04:03.156873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:57.911 BaseBdev3 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.911 [ 00:14:57.911 { 00:14:57.911 "name": "BaseBdev3", 00:14:57.911 "aliases": [ 00:14:57.911 "30a8b9c9-dc44-4c69-b010-d26b955ab74a" 00:14:57.911 ], 00:14:57.911 "product_name": "Malloc disk", 00:14:57.911 "block_size": 512, 00:14:57.911 "num_blocks": 65536, 00:14:57.911 "uuid": "30a8b9c9-dc44-4c69-b010-d26b955ab74a", 00:14:57.911 "assigned_rate_limits": { 00:14:57.911 "rw_ios_per_sec": 0, 00:14:57.911 "rw_mbytes_per_sec": 0, 00:14:57.911 "r_mbytes_per_sec": 0, 00:14:57.911 "w_mbytes_per_sec": 0 00:14:57.911 }, 00:14:57.911 "claimed": true, 00:14:57.911 "claim_type": "exclusive_write", 00:14:57.911 "zoned": false, 00:14:57.911 "supported_io_types": { 00:14:57.911 "read": true, 00:14:57.911 "write": true, 00:14:57.911 "unmap": true, 00:14:57.911 "flush": true, 00:14:57.911 "reset": true, 00:14:57.911 "nvme_admin": false, 00:14:57.911 "nvme_io": false, 00:14:57.911 "nvme_io_md": false, 00:14:57.911 "write_zeroes": true, 00:14:57.911 "zcopy": true, 00:14:57.911 "get_zone_info": false, 00:14:57.911 "zone_management": false, 00:14:57.911 "zone_append": false, 00:14:57.911 "compare": false, 00:14:57.911 "compare_and_write": false, 00:14:57.911 "abort": true, 00:14:57.911 "seek_hole": false, 00:14:57.911 "seek_data": false, 00:14:57.911 "copy": true, 00:14:57.911 "nvme_iov_md": false 00:14:57.911 }, 00:14:57.911 "memory_domains": [ 00:14:57.911 { 00:14:57.911 "dma_device_id": "system", 00:14:57.911 "dma_device_type": 1 00:14:57.911 }, 00:14:57.911 { 00:14:57.911 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:57.911 "dma_device_type": 2 00:14:57.911 } 00:14:57.911 ], 00:14:57.911 "driver_specific": {} 00:14:57.911 } 00:14:57.911 ] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.911 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.911 "name": "Existed_Raid", 00:14:57.911 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:57.911 "strip_size_kb": 64, 00:14:57.911 "state": "configuring", 00:14:57.911 "raid_level": "raid5f", 00:14:57.911 "superblock": true, 00:14:57.911 "num_base_bdevs": 4, 00:14:57.911 "num_base_bdevs_discovered": 3, 00:14:57.911 "num_base_bdevs_operational": 4, 00:14:57.911 "base_bdevs_list": [ 00:14:57.911 { 00:14:57.911 "name": "BaseBdev1", 00:14:57.911 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:57.911 "is_configured": true, 00:14:57.911 "data_offset": 2048, 00:14:57.911 "data_size": 63488 00:14:57.911 }, 00:14:57.911 { 00:14:57.911 "name": "BaseBdev2", 00:14:57.911 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:57.911 "is_configured": true, 00:14:57.911 "data_offset": 2048, 00:14:57.911 "data_size": 63488 00:14:57.911 }, 00:14:57.911 { 00:14:57.911 "name": "BaseBdev3", 00:14:57.911 "uuid": "30a8b9c9-dc44-4c69-b010-d26b955ab74a", 00:14:57.911 "is_configured": true, 00:14:57.911 "data_offset": 2048, 00:14:57.911 "data_size": 63488 00:14:57.911 }, 00:14:57.912 { 00:14:57.912 "name": "BaseBdev4", 00:14:57.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.912 "is_configured": false, 00:14:57.912 "data_offset": 0, 00:14:57.912 "data_size": 0 00:14:57.912 } 00:14:57.912 ] 00:14:57.912 }' 00:14:57.912 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.912 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.172 [2024-10-29 11:04:03.659283] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:58.172 [2024-10-29 11:04:03.659660] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:14:58.172 [2024-10-29 11:04:03.659720] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:58.172 BaseBdev4 00:14:58.172 [2024-10-29 11:04:03.660058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:14:58.172 [2024-10-29 11:04:03.660593] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:14:58.172 [2024-10-29 11:04:03.660618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:14:58.172 [2024-10-29 11:04:03.660746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.172 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.433 [ 00:14:58.433 { 00:14:58.433 "name": "BaseBdev4", 00:14:58.433 "aliases": [ 00:14:58.433 "1714a8bf-f790-4a74-ac68-b1bdbcae4f47" 00:14:58.433 ], 00:14:58.433 "product_name": "Malloc disk", 00:14:58.433 "block_size": 512, 00:14:58.433 "num_blocks": 65536, 00:14:58.433 "uuid": "1714a8bf-f790-4a74-ac68-b1bdbcae4f47", 00:14:58.433 "assigned_rate_limits": { 00:14:58.433 "rw_ios_per_sec": 0, 00:14:58.433 "rw_mbytes_per_sec": 0, 00:14:58.433 "r_mbytes_per_sec": 0, 00:14:58.433 "w_mbytes_per_sec": 0 00:14:58.433 }, 00:14:58.433 "claimed": true, 00:14:58.433 "claim_type": "exclusive_write", 00:14:58.433 "zoned": false, 00:14:58.433 "supported_io_types": { 00:14:58.433 "read": true, 00:14:58.433 "write": true, 00:14:58.433 "unmap": true, 00:14:58.433 "flush": true, 00:14:58.433 "reset": true, 00:14:58.433 "nvme_admin": false, 00:14:58.433 "nvme_io": false, 00:14:58.433 "nvme_io_md": false, 00:14:58.433 "write_zeroes": true, 00:14:58.433 "zcopy": true, 00:14:58.433 "get_zone_info": false, 00:14:58.433 "zone_management": false, 00:14:58.433 "zone_append": false, 00:14:58.433 "compare": false, 00:14:58.433 "compare_and_write": false, 00:14:58.433 "abort": true, 00:14:58.433 "seek_hole": false, 00:14:58.433 "seek_data": false, 00:14:58.433 "copy": true, 00:14:58.433 "nvme_iov_md": false 00:14:58.433 }, 00:14:58.433 "memory_domains": [ 00:14:58.433 { 00:14:58.433 "dma_device_id": "system", 00:14:58.433 "dma_device_type": 1 00:14:58.433 }, 00:14:58.433 { 00:14:58.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:58.433 "dma_device_type": 2 00:14:58.433 } 00:14:58.433 ], 00:14:58.433 "driver_specific": {} 00:14:58.433 } 00:14:58.433 ] 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.433 "name": "Existed_Raid", 00:14:58.433 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:58.433 "strip_size_kb": 64, 00:14:58.433 "state": "online", 00:14:58.433 "raid_level": "raid5f", 00:14:58.433 "superblock": true, 00:14:58.433 "num_base_bdevs": 4, 00:14:58.433 "num_base_bdevs_discovered": 4, 00:14:58.433 "num_base_bdevs_operational": 4, 00:14:58.433 "base_bdevs_list": [ 00:14:58.433 { 00:14:58.433 "name": "BaseBdev1", 00:14:58.433 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:58.433 "is_configured": true, 00:14:58.433 "data_offset": 2048, 00:14:58.433 "data_size": 63488 00:14:58.433 }, 00:14:58.433 { 00:14:58.433 "name": "BaseBdev2", 00:14:58.433 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:58.433 "is_configured": true, 00:14:58.433 "data_offset": 2048, 00:14:58.433 "data_size": 63488 00:14:58.433 }, 00:14:58.433 { 00:14:58.433 "name": "BaseBdev3", 00:14:58.433 "uuid": "30a8b9c9-dc44-4c69-b010-d26b955ab74a", 00:14:58.433 "is_configured": true, 00:14:58.433 "data_offset": 2048, 00:14:58.433 "data_size": 63488 00:14:58.433 }, 00:14:58.433 { 00:14:58.433 "name": "BaseBdev4", 00:14:58.433 "uuid": "1714a8bf-f790-4a74-ac68-b1bdbcae4f47", 00:14:58.433 "is_configured": true, 00:14:58.433 "data_offset": 2048, 00:14:58.433 "data_size": 63488 00:14:58.433 } 00:14:58.433 ] 00:14:58.433 }' 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.433 11:04:03 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:58.694 [2024-10-29 11:04:04.174686] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:58.694 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.955 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:58.955 "name": "Existed_Raid", 00:14:58.955 "aliases": [ 00:14:58.955 "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9" 00:14:58.955 ], 00:14:58.955 "product_name": "Raid Volume", 00:14:58.955 "block_size": 512, 00:14:58.955 "num_blocks": 190464, 00:14:58.955 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:58.955 "assigned_rate_limits": { 00:14:58.955 "rw_ios_per_sec": 0, 00:14:58.955 "rw_mbytes_per_sec": 0, 00:14:58.955 "r_mbytes_per_sec": 0, 00:14:58.955 "w_mbytes_per_sec": 0 00:14:58.955 }, 00:14:58.955 "claimed": false, 00:14:58.955 "zoned": false, 00:14:58.955 "supported_io_types": { 00:14:58.955 "read": true, 00:14:58.955 "write": true, 00:14:58.955 "unmap": false, 00:14:58.955 "flush": false, 00:14:58.955 "reset": true, 00:14:58.955 "nvme_admin": false, 00:14:58.955 "nvme_io": false, 00:14:58.955 "nvme_io_md": false, 00:14:58.955 "write_zeroes": true, 00:14:58.955 "zcopy": false, 00:14:58.955 "get_zone_info": false, 00:14:58.955 "zone_management": false, 00:14:58.955 "zone_append": false, 00:14:58.955 "compare": false, 00:14:58.955 "compare_and_write": false, 00:14:58.955 "abort": false, 00:14:58.955 "seek_hole": false, 00:14:58.955 "seek_data": false, 00:14:58.955 "copy": false, 00:14:58.955 "nvme_iov_md": false 00:14:58.955 }, 00:14:58.955 "driver_specific": { 00:14:58.955 "raid": { 00:14:58.955 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:58.955 "strip_size_kb": 64, 00:14:58.955 "state": "online", 00:14:58.955 "raid_level": "raid5f", 00:14:58.955 "superblock": true, 00:14:58.955 "num_base_bdevs": 4, 00:14:58.955 "num_base_bdevs_discovered": 4, 00:14:58.955 "num_base_bdevs_operational": 4, 00:14:58.955 "base_bdevs_list": [ 00:14:58.955 { 00:14:58.955 "name": "BaseBdev1", 00:14:58.955 "uuid": "6080e89f-b6a9-4746-a388-0fa619605eb6", 00:14:58.955 "is_configured": true, 00:14:58.955 "data_offset": 2048, 00:14:58.955 "data_size": 63488 00:14:58.955 }, 00:14:58.955 { 00:14:58.955 "name": "BaseBdev2", 00:14:58.955 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:58.955 "is_configured": true, 00:14:58.955 "data_offset": 2048, 00:14:58.955 "data_size": 63488 00:14:58.955 }, 00:14:58.955 { 00:14:58.955 "name": "BaseBdev3", 00:14:58.955 "uuid": "30a8b9c9-dc44-4c69-b010-d26b955ab74a", 00:14:58.955 "is_configured": true, 00:14:58.955 "data_offset": 2048, 00:14:58.955 "data_size": 63488 00:14:58.955 }, 00:14:58.955 { 00:14:58.955 "name": "BaseBdev4", 00:14:58.955 "uuid": "1714a8bf-f790-4a74-ac68-b1bdbcae4f47", 00:14:58.956 "is_configured": true, 00:14:58.956 "data_offset": 2048, 00:14:58.956 "data_size": 63488 00:14:58.956 } 00:14:58.956 ] 00:14:58.956 } 00:14:58.956 } 00:14:58.956 }' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:58.956 BaseBdev2 00:14:58.956 BaseBdev3 00:14:58.956 BaseBdev4' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.956 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.216 [2024-10-29 11:04:04.473975] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.216 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:59.216 "name": "Existed_Raid", 00:14:59.216 "uuid": "6f1e33f6-87cc-4f9b-a2ed-fbc648983ed9", 00:14:59.216 "strip_size_kb": 64, 00:14:59.216 "state": "online", 00:14:59.216 "raid_level": "raid5f", 00:14:59.216 "superblock": true, 00:14:59.216 "num_base_bdevs": 4, 00:14:59.216 "num_base_bdevs_discovered": 3, 00:14:59.216 "num_base_bdevs_operational": 3, 00:14:59.216 "base_bdevs_list": [ 00:14:59.216 { 00:14:59.216 "name": null, 00:14:59.216 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:59.216 "is_configured": false, 00:14:59.216 "data_offset": 0, 00:14:59.217 "data_size": 63488 00:14:59.217 }, 00:14:59.217 { 00:14:59.217 "name": "BaseBdev2", 00:14:59.217 "uuid": "23ab6273-bd28-4dee-b90c-cced9b0fe468", 00:14:59.217 "is_configured": true, 00:14:59.217 "data_offset": 2048, 00:14:59.217 "data_size": 63488 00:14:59.217 }, 00:14:59.217 { 00:14:59.217 "name": "BaseBdev3", 00:14:59.217 "uuid": "30a8b9c9-dc44-4c69-b010-d26b955ab74a", 00:14:59.217 "is_configured": true, 00:14:59.217 "data_offset": 2048, 00:14:59.217 "data_size": 63488 00:14:59.217 }, 00:14:59.217 { 00:14:59.217 "name": "BaseBdev4", 00:14:59.217 "uuid": "1714a8bf-f790-4a74-ac68-b1bdbcae4f47", 00:14:59.217 "is_configured": true, 00:14:59.217 "data_offset": 2048, 00:14:59.217 "data_size": 63488 00:14:59.217 } 00:14:59.217 ] 00:14:59.217 }' 00:14:59.217 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:59.217 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.476 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:59.477 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:59.737 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:59.737 11:04:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:59.737 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 [2024-10-29 11:04:04.988777] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:59.737 [2024-10-29 11:04:04.988928] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:59.737 [2024-10-29 11:04:05.000267] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 [2024-10-29 11:04:05.060184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 [2024-10-29 11:04:05.131239] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:59.737 [2024-10-29 11:04:05.131289] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:59.737 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.738 BaseBdev2 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.738 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.738 [ 00:14:59.738 { 00:14:59.738 "name": "BaseBdev2", 00:14:59.738 "aliases": [ 00:14:59.998 "a91a1975-222f-4bd5-9ea2-7ad33184546d" 00:14:59.998 ], 00:14:59.998 "product_name": "Malloc disk", 00:14:59.998 "block_size": 512, 00:14:59.998 "num_blocks": 65536, 00:14:59.998 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:14:59.998 "assigned_rate_limits": { 00:14:59.998 "rw_ios_per_sec": 0, 00:14:59.998 "rw_mbytes_per_sec": 0, 00:14:59.998 "r_mbytes_per_sec": 0, 00:14:59.998 "w_mbytes_per_sec": 0 00:14:59.998 }, 00:14:59.998 "claimed": false, 00:14:59.998 "zoned": false, 00:14:59.998 "supported_io_types": { 00:14:59.998 "read": true, 00:14:59.998 "write": true, 00:14:59.998 "unmap": true, 00:14:59.998 "flush": true, 00:14:59.998 "reset": true, 00:14:59.998 "nvme_admin": false, 00:14:59.998 "nvme_io": false, 00:14:59.998 "nvme_io_md": false, 00:14:59.998 "write_zeroes": true, 00:14:59.998 "zcopy": true, 00:14:59.998 "get_zone_info": false, 00:14:59.998 "zone_management": false, 00:14:59.998 "zone_append": false, 00:14:59.998 "compare": false, 00:14:59.998 "compare_and_write": false, 00:14:59.998 "abort": true, 00:14:59.998 "seek_hole": false, 00:14:59.998 "seek_data": false, 00:14:59.998 "copy": true, 00:14:59.998 "nvme_iov_md": false 00:14:59.998 }, 00:14:59.998 "memory_domains": [ 00:14:59.998 { 00:14:59.999 "dma_device_id": "system", 00:14:59.999 "dma_device_type": 1 00:14:59.999 }, 00:14:59.999 { 00:14:59.999 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:59.999 "dma_device_type": 2 00:14:59.999 } 00:14:59.999 ], 00:14:59.999 "driver_specific": {} 00:14:59.999 } 00:14:59.999 ] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 BaseBdev3 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev3 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 [ 00:14:59.999 { 00:14:59.999 "name": "BaseBdev3", 00:14:59.999 "aliases": [ 00:14:59.999 "d0a268f6-25a5-457f-8bcf-a9b560302e8f" 00:14:59.999 ], 00:14:59.999 "product_name": "Malloc disk", 00:14:59.999 "block_size": 512, 00:14:59.999 "num_blocks": 65536, 00:14:59.999 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:14:59.999 "assigned_rate_limits": { 00:14:59.999 "rw_ios_per_sec": 0, 00:14:59.999 "rw_mbytes_per_sec": 0, 00:14:59.999 "r_mbytes_per_sec": 0, 00:14:59.999 "w_mbytes_per_sec": 0 00:14:59.999 }, 00:14:59.999 "claimed": false, 00:14:59.999 "zoned": false, 00:14:59.999 "supported_io_types": { 00:14:59.999 "read": true, 00:14:59.999 "write": true, 00:14:59.999 "unmap": true, 00:14:59.999 "flush": true, 00:14:59.999 "reset": true, 00:14:59.999 "nvme_admin": false, 00:14:59.999 "nvme_io": false, 00:14:59.999 "nvme_io_md": false, 00:14:59.999 "write_zeroes": true, 00:14:59.999 "zcopy": true, 00:14:59.999 "get_zone_info": false, 00:14:59.999 "zone_management": false, 00:14:59.999 "zone_append": false, 00:14:59.999 "compare": false, 00:14:59.999 "compare_and_write": false, 00:14:59.999 "abort": true, 00:14:59.999 "seek_hole": false, 00:14:59.999 "seek_data": false, 00:14:59.999 "copy": true, 00:14:59.999 "nvme_iov_md": false 00:14:59.999 }, 00:14:59.999 "memory_domains": [ 00:14:59.999 { 00:14:59.999 "dma_device_id": "system", 00:14:59.999 "dma_device_type": 1 00:14:59.999 }, 00:14:59.999 { 00:14:59.999 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:59.999 "dma_device_type": 2 00:14:59.999 } 00:14:59.999 ], 00:14:59.999 "driver_specific": {} 00:14:59.999 } 00:14:59.999 ] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 BaseBdev4 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev4 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.999 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.999 [ 00:14:59.999 { 00:14:59.999 "name": "BaseBdev4", 00:14:59.999 "aliases": [ 00:14:59.999 "441e9fb7-5c8f-4c75-beac-3ca4e6d27630" 00:14:59.999 ], 00:14:59.999 "product_name": "Malloc disk", 00:14:59.999 "block_size": 512, 00:14:59.999 "num_blocks": 65536, 00:14:59.999 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:14:59.999 "assigned_rate_limits": { 00:14:59.999 "rw_ios_per_sec": 0, 00:14:59.999 "rw_mbytes_per_sec": 0, 00:14:59.999 "r_mbytes_per_sec": 0, 00:14:59.999 "w_mbytes_per_sec": 0 00:14:59.999 }, 00:14:59.999 "claimed": false, 00:14:59.999 "zoned": false, 00:14:59.999 "supported_io_types": { 00:14:59.999 "read": true, 00:14:59.999 "write": true, 00:14:59.999 "unmap": true, 00:14:59.999 "flush": true, 00:14:59.999 "reset": true, 00:14:59.999 "nvme_admin": false, 00:14:59.999 "nvme_io": false, 00:14:59.999 "nvme_io_md": false, 00:14:59.999 "write_zeroes": true, 00:14:59.999 "zcopy": true, 00:14:59.999 "get_zone_info": false, 00:14:59.999 "zone_management": false, 00:14:59.999 "zone_append": false, 00:14:59.999 "compare": false, 00:14:59.999 "compare_and_write": false, 00:14:59.999 "abort": true, 00:14:59.999 "seek_hole": false, 00:14:59.999 "seek_data": false, 00:14:59.999 "copy": true, 00:14:59.999 "nvme_iov_md": false 00:14:59.999 }, 00:14:59.999 "memory_domains": [ 00:14:59.999 { 00:14:59.999 "dma_device_id": "system", 00:14:59.999 "dma_device_type": 1 00:14:59.999 }, 00:14:59.999 { 00:14:59.999 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:00.000 "dma_device_type": 2 00:15:00.000 } 00:15:00.000 ], 00:15:00.000 "driver_specific": {} 00:15:00.000 } 00:15:00.000 ] 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.000 [2024-10-29 11:04:05.360345] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:00.000 [2024-10-29 11:04:05.360450] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:00.000 [2024-10-29 11:04:05.360535] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:00.000 [2024-10-29 11:04:05.362397] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:00.000 [2024-10-29 11:04:05.362496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.000 "name": "Existed_Raid", 00:15:00.000 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:00.000 "strip_size_kb": 64, 00:15:00.000 "state": "configuring", 00:15:00.000 "raid_level": "raid5f", 00:15:00.000 "superblock": true, 00:15:00.000 "num_base_bdevs": 4, 00:15:00.000 "num_base_bdevs_discovered": 3, 00:15:00.000 "num_base_bdevs_operational": 4, 00:15:00.000 "base_bdevs_list": [ 00:15:00.000 { 00:15:00.000 "name": "BaseBdev1", 00:15:00.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.000 "is_configured": false, 00:15:00.000 "data_offset": 0, 00:15:00.000 "data_size": 0 00:15:00.000 }, 00:15:00.000 { 00:15:00.000 "name": "BaseBdev2", 00:15:00.000 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:00.000 "is_configured": true, 00:15:00.000 "data_offset": 2048, 00:15:00.000 "data_size": 63488 00:15:00.000 }, 00:15:00.000 { 00:15:00.000 "name": "BaseBdev3", 00:15:00.000 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:00.000 "is_configured": true, 00:15:00.000 "data_offset": 2048, 00:15:00.000 "data_size": 63488 00:15:00.000 }, 00:15:00.000 { 00:15:00.000 "name": "BaseBdev4", 00:15:00.000 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:00.000 "is_configured": true, 00:15:00.000 "data_offset": 2048, 00:15:00.000 "data_size": 63488 00:15:00.000 } 00:15:00.000 ] 00:15:00.000 }' 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.000 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.570 [2024-10-29 11:04:05.820000] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:00.570 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.571 "name": "Existed_Raid", 00:15:00.571 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:00.571 "strip_size_kb": 64, 00:15:00.571 "state": "configuring", 00:15:00.571 "raid_level": "raid5f", 00:15:00.571 "superblock": true, 00:15:00.571 "num_base_bdevs": 4, 00:15:00.571 "num_base_bdevs_discovered": 2, 00:15:00.571 "num_base_bdevs_operational": 4, 00:15:00.571 "base_bdevs_list": [ 00:15:00.571 { 00:15:00.571 "name": "BaseBdev1", 00:15:00.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.571 "is_configured": false, 00:15:00.571 "data_offset": 0, 00:15:00.571 "data_size": 0 00:15:00.571 }, 00:15:00.571 { 00:15:00.571 "name": null, 00:15:00.571 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:00.571 "is_configured": false, 00:15:00.571 "data_offset": 0, 00:15:00.571 "data_size": 63488 00:15:00.571 }, 00:15:00.571 { 00:15:00.571 "name": "BaseBdev3", 00:15:00.571 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:00.571 "is_configured": true, 00:15:00.571 "data_offset": 2048, 00:15:00.571 "data_size": 63488 00:15:00.571 }, 00:15:00.571 { 00:15:00.571 "name": "BaseBdev4", 00:15:00.571 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:00.571 "is_configured": true, 00:15:00.571 "data_offset": 2048, 00:15:00.571 "data_size": 63488 00:15:00.571 } 00:15:00.571 ] 00:15:00.571 }' 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.571 11:04:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.831 [2024-10-29 11:04:06.306557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:00.831 BaseBdev1 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:15:00.831 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.832 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.092 [ 00:15:01.092 { 00:15:01.092 "name": "BaseBdev1", 00:15:01.092 "aliases": [ 00:15:01.092 "96da32bf-9504-4923-88f2-61267db13aea" 00:15:01.092 ], 00:15:01.092 "product_name": "Malloc disk", 00:15:01.092 "block_size": 512, 00:15:01.092 "num_blocks": 65536, 00:15:01.092 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:01.092 "assigned_rate_limits": { 00:15:01.092 "rw_ios_per_sec": 0, 00:15:01.092 "rw_mbytes_per_sec": 0, 00:15:01.092 "r_mbytes_per_sec": 0, 00:15:01.092 "w_mbytes_per_sec": 0 00:15:01.092 }, 00:15:01.092 "claimed": true, 00:15:01.092 "claim_type": "exclusive_write", 00:15:01.092 "zoned": false, 00:15:01.092 "supported_io_types": { 00:15:01.092 "read": true, 00:15:01.092 "write": true, 00:15:01.092 "unmap": true, 00:15:01.092 "flush": true, 00:15:01.092 "reset": true, 00:15:01.092 "nvme_admin": false, 00:15:01.092 "nvme_io": false, 00:15:01.092 "nvme_io_md": false, 00:15:01.092 "write_zeroes": true, 00:15:01.092 "zcopy": true, 00:15:01.092 "get_zone_info": false, 00:15:01.092 "zone_management": false, 00:15:01.092 "zone_append": false, 00:15:01.092 "compare": false, 00:15:01.092 "compare_and_write": false, 00:15:01.092 "abort": true, 00:15:01.092 "seek_hole": false, 00:15:01.092 "seek_data": false, 00:15:01.092 "copy": true, 00:15:01.092 "nvme_iov_md": false 00:15:01.092 }, 00:15:01.092 "memory_domains": [ 00:15:01.092 { 00:15:01.092 "dma_device_id": "system", 00:15:01.092 "dma_device_type": 1 00:15:01.092 }, 00:15:01.092 { 00:15:01.092 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:01.092 "dma_device_type": 2 00:15:01.092 } 00:15:01.092 ], 00:15:01.092 "driver_specific": {} 00:15:01.092 } 00:15:01.092 ] 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.092 "name": "Existed_Raid", 00:15:01.092 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:01.092 "strip_size_kb": 64, 00:15:01.092 "state": "configuring", 00:15:01.092 "raid_level": "raid5f", 00:15:01.092 "superblock": true, 00:15:01.092 "num_base_bdevs": 4, 00:15:01.092 "num_base_bdevs_discovered": 3, 00:15:01.092 "num_base_bdevs_operational": 4, 00:15:01.092 "base_bdevs_list": [ 00:15:01.092 { 00:15:01.092 "name": "BaseBdev1", 00:15:01.092 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:01.092 "is_configured": true, 00:15:01.092 "data_offset": 2048, 00:15:01.092 "data_size": 63488 00:15:01.092 }, 00:15:01.092 { 00:15:01.092 "name": null, 00:15:01.092 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:01.092 "is_configured": false, 00:15:01.092 "data_offset": 0, 00:15:01.092 "data_size": 63488 00:15:01.092 }, 00:15:01.092 { 00:15:01.092 "name": "BaseBdev3", 00:15:01.092 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:01.092 "is_configured": true, 00:15:01.092 "data_offset": 2048, 00:15:01.092 "data_size": 63488 00:15:01.092 }, 00:15:01.092 { 00:15:01.092 "name": "BaseBdev4", 00:15:01.092 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:01.092 "is_configured": true, 00:15:01.092 "data_offset": 2048, 00:15:01.092 "data_size": 63488 00:15:01.092 } 00:15:01.092 ] 00:15:01.092 }' 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.092 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.352 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.613 [2024-10-29 11:04:06.853681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.613 "name": "Existed_Raid", 00:15:01.613 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:01.613 "strip_size_kb": 64, 00:15:01.613 "state": "configuring", 00:15:01.613 "raid_level": "raid5f", 00:15:01.613 "superblock": true, 00:15:01.613 "num_base_bdevs": 4, 00:15:01.613 "num_base_bdevs_discovered": 2, 00:15:01.613 "num_base_bdevs_operational": 4, 00:15:01.613 "base_bdevs_list": [ 00:15:01.613 { 00:15:01.613 "name": "BaseBdev1", 00:15:01.613 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:01.613 "is_configured": true, 00:15:01.613 "data_offset": 2048, 00:15:01.613 "data_size": 63488 00:15:01.613 }, 00:15:01.613 { 00:15:01.613 "name": null, 00:15:01.613 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:01.613 "is_configured": false, 00:15:01.613 "data_offset": 0, 00:15:01.613 "data_size": 63488 00:15:01.613 }, 00:15:01.613 { 00:15:01.613 "name": null, 00:15:01.613 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:01.613 "is_configured": false, 00:15:01.613 "data_offset": 0, 00:15:01.613 "data_size": 63488 00:15:01.613 }, 00:15:01.613 { 00:15:01.613 "name": "BaseBdev4", 00:15:01.613 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:01.613 "is_configured": true, 00:15:01.613 "data_offset": 2048, 00:15:01.613 "data_size": 63488 00:15:01.613 } 00:15:01.613 ] 00:15:01.613 }' 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.613 11:04:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.874 [2024-10-29 11:04:07.356901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.874 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.134 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.134 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.134 "name": "Existed_Raid", 00:15:02.134 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:02.134 "strip_size_kb": 64, 00:15:02.134 "state": "configuring", 00:15:02.134 "raid_level": "raid5f", 00:15:02.134 "superblock": true, 00:15:02.134 "num_base_bdevs": 4, 00:15:02.134 "num_base_bdevs_discovered": 3, 00:15:02.134 "num_base_bdevs_operational": 4, 00:15:02.134 "base_bdevs_list": [ 00:15:02.134 { 00:15:02.134 "name": "BaseBdev1", 00:15:02.134 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:02.134 "is_configured": true, 00:15:02.134 "data_offset": 2048, 00:15:02.134 "data_size": 63488 00:15:02.134 }, 00:15:02.134 { 00:15:02.134 "name": null, 00:15:02.134 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:02.134 "is_configured": false, 00:15:02.134 "data_offset": 0, 00:15:02.134 "data_size": 63488 00:15:02.134 }, 00:15:02.134 { 00:15:02.134 "name": "BaseBdev3", 00:15:02.134 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:02.134 "is_configured": true, 00:15:02.134 "data_offset": 2048, 00:15:02.134 "data_size": 63488 00:15:02.134 }, 00:15:02.134 { 00:15:02.134 "name": "BaseBdev4", 00:15:02.134 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:02.134 "is_configured": true, 00:15:02.134 "data_offset": 2048, 00:15:02.134 "data_size": 63488 00:15:02.134 } 00:15:02.134 ] 00:15:02.134 }' 00:15:02.134 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.134 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.395 [2024-10-29 11:04:07.860067] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.395 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.656 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.656 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.656 "name": "Existed_Raid", 00:15:02.656 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:02.656 "strip_size_kb": 64, 00:15:02.656 "state": "configuring", 00:15:02.656 "raid_level": "raid5f", 00:15:02.656 "superblock": true, 00:15:02.656 "num_base_bdevs": 4, 00:15:02.656 "num_base_bdevs_discovered": 2, 00:15:02.656 "num_base_bdevs_operational": 4, 00:15:02.656 "base_bdevs_list": [ 00:15:02.656 { 00:15:02.656 "name": null, 00:15:02.656 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:02.656 "is_configured": false, 00:15:02.656 "data_offset": 0, 00:15:02.656 "data_size": 63488 00:15:02.656 }, 00:15:02.656 { 00:15:02.656 "name": null, 00:15:02.656 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:02.656 "is_configured": false, 00:15:02.656 "data_offset": 0, 00:15:02.656 "data_size": 63488 00:15:02.656 }, 00:15:02.656 { 00:15:02.656 "name": "BaseBdev3", 00:15:02.656 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:02.656 "is_configured": true, 00:15:02.656 "data_offset": 2048, 00:15:02.656 "data_size": 63488 00:15:02.656 }, 00:15:02.656 { 00:15:02.656 "name": "BaseBdev4", 00:15:02.656 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:02.656 "is_configured": true, 00:15:02.656 "data_offset": 2048, 00:15:02.656 "data_size": 63488 00:15:02.656 } 00:15:02.656 ] 00:15:02.656 }' 00:15:02.656 11:04:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.656 11:04:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.916 [2024-10-29 11:04:08.389903] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.916 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.177 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.177 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.177 "name": "Existed_Raid", 00:15:03.177 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:03.177 "strip_size_kb": 64, 00:15:03.177 "state": "configuring", 00:15:03.177 "raid_level": "raid5f", 00:15:03.177 "superblock": true, 00:15:03.177 "num_base_bdevs": 4, 00:15:03.177 "num_base_bdevs_discovered": 3, 00:15:03.177 "num_base_bdevs_operational": 4, 00:15:03.177 "base_bdevs_list": [ 00:15:03.177 { 00:15:03.177 "name": null, 00:15:03.177 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:03.177 "is_configured": false, 00:15:03.177 "data_offset": 0, 00:15:03.177 "data_size": 63488 00:15:03.177 }, 00:15:03.177 { 00:15:03.177 "name": "BaseBdev2", 00:15:03.177 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:03.177 "is_configured": true, 00:15:03.177 "data_offset": 2048, 00:15:03.177 "data_size": 63488 00:15:03.177 }, 00:15:03.177 { 00:15:03.177 "name": "BaseBdev3", 00:15:03.177 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:03.177 "is_configured": true, 00:15:03.177 "data_offset": 2048, 00:15:03.177 "data_size": 63488 00:15:03.177 }, 00:15:03.177 { 00:15:03.177 "name": "BaseBdev4", 00:15:03.177 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:03.177 "is_configured": true, 00:15:03.177 "data_offset": 2048, 00:15:03.177 "data_size": 63488 00:15:03.177 } 00:15:03.177 ] 00:15:03.177 }' 00:15:03.177 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.177 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 96da32bf-9504-4923-88f2-61267db13aea 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.438 NewBaseBdev 00:15:03.438 [2024-10-29 11:04:08.923905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:03.438 [2024-10-29 11:04:08.924118] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:15:03.438 [2024-10-29 11:04:08.924132] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:03.438 [2024-10-29 11:04:08.924401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:15:03.438 [2024-10-29 11:04:08.924913] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:15:03.438 [2024-10-29 11:04:08.924990] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006d00 00:15:03.438 [2024-10-29 11:04:08.925118] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local bdev_name=NewBaseBdev 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local i 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.438 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.700 [ 00:15:03.700 { 00:15:03.700 "name": "NewBaseBdev", 00:15:03.700 "aliases": [ 00:15:03.700 "96da32bf-9504-4923-88f2-61267db13aea" 00:15:03.700 ], 00:15:03.700 "product_name": "Malloc disk", 00:15:03.700 "block_size": 512, 00:15:03.700 "num_blocks": 65536, 00:15:03.700 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:03.700 "assigned_rate_limits": { 00:15:03.700 "rw_ios_per_sec": 0, 00:15:03.700 "rw_mbytes_per_sec": 0, 00:15:03.700 "r_mbytes_per_sec": 0, 00:15:03.700 "w_mbytes_per_sec": 0 00:15:03.700 }, 00:15:03.700 "claimed": true, 00:15:03.700 "claim_type": "exclusive_write", 00:15:03.700 "zoned": false, 00:15:03.700 "supported_io_types": { 00:15:03.700 "read": true, 00:15:03.700 "write": true, 00:15:03.700 "unmap": true, 00:15:03.700 "flush": true, 00:15:03.700 "reset": true, 00:15:03.700 "nvme_admin": false, 00:15:03.700 "nvme_io": false, 00:15:03.700 "nvme_io_md": false, 00:15:03.700 "write_zeroes": true, 00:15:03.700 "zcopy": true, 00:15:03.700 "get_zone_info": false, 00:15:03.700 "zone_management": false, 00:15:03.700 "zone_append": false, 00:15:03.700 "compare": false, 00:15:03.700 "compare_and_write": false, 00:15:03.700 "abort": true, 00:15:03.700 "seek_hole": false, 00:15:03.700 "seek_data": false, 00:15:03.700 "copy": true, 00:15:03.700 "nvme_iov_md": false 00:15:03.700 }, 00:15:03.700 "memory_domains": [ 00:15:03.700 { 00:15:03.700 "dma_device_id": "system", 00:15:03.700 "dma_device_type": 1 00:15:03.700 }, 00:15:03.700 { 00:15:03.700 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:03.700 "dma_device_type": 2 00:15:03.700 } 00:15:03.700 ], 00:15:03.700 "driver_specific": {} 00:15:03.700 } 00:15:03.700 ] 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@909 -- # return 0 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.700 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.700 "name": "Existed_Raid", 00:15:03.700 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:03.700 "strip_size_kb": 64, 00:15:03.700 "state": "online", 00:15:03.700 "raid_level": "raid5f", 00:15:03.700 "superblock": true, 00:15:03.700 "num_base_bdevs": 4, 00:15:03.700 "num_base_bdevs_discovered": 4, 00:15:03.701 "num_base_bdevs_operational": 4, 00:15:03.701 "base_bdevs_list": [ 00:15:03.701 { 00:15:03.701 "name": "NewBaseBdev", 00:15:03.701 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:03.701 "is_configured": true, 00:15:03.701 "data_offset": 2048, 00:15:03.701 "data_size": 63488 00:15:03.701 }, 00:15:03.701 { 00:15:03.701 "name": "BaseBdev2", 00:15:03.701 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:03.701 "is_configured": true, 00:15:03.701 "data_offset": 2048, 00:15:03.701 "data_size": 63488 00:15:03.701 }, 00:15:03.701 { 00:15:03.701 "name": "BaseBdev3", 00:15:03.701 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:03.701 "is_configured": true, 00:15:03.701 "data_offset": 2048, 00:15:03.701 "data_size": 63488 00:15:03.701 }, 00:15:03.701 { 00:15:03.701 "name": "BaseBdev4", 00:15:03.701 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:03.701 "is_configured": true, 00:15:03.701 "data_offset": 2048, 00:15:03.701 "data_size": 63488 00:15:03.701 } 00:15:03.701 ] 00:15:03.701 }' 00:15:03.701 11:04:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.701 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.961 [2024-10-29 11:04:09.383410] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.961 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:03.961 "name": "Existed_Raid", 00:15:03.961 "aliases": [ 00:15:03.961 "cf215e54-a9e3-4ce9-9197-3d4f3f032c20" 00:15:03.962 ], 00:15:03.962 "product_name": "Raid Volume", 00:15:03.962 "block_size": 512, 00:15:03.962 "num_blocks": 190464, 00:15:03.962 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:03.962 "assigned_rate_limits": { 00:15:03.962 "rw_ios_per_sec": 0, 00:15:03.962 "rw_mbytes_per_sec": 0, 00:15:03.962 "r_mbytes_per_sec": 0, 00:15:03.962 "w_mbytes_per_sec": 0 00:15:03.962 }, 00:15:03.962 "claimed": false, 00:15:03.962 "zoned": false, 00:15:03.962 "supported_io_types": { 00:15:03.962 "read": true, 00:15:03.962 "write": true, 00:15:03.962 "unmap": false, 00:15:03.962 "flush": false, 00:15:03.962 "reset": true, 00:15:03.962 "nvme_admin": false, 00:15:03.962 "nvme_io": false, 00:15:03.962 "nvme_io_md": false, 00:15:03.962 "write_zeroes": true, 00:15:03.962 "zcopy": false, 00:15:03.962 "get_zone_info": false, 00:15:03.962 "zone_management": false, 00:15:03.962 "zone_append": false, 00:15:03.962 "compare": false, 00:15:03.962 "compare_and_write": false, 00:15:03.962 "abort": false, 00:15:03.962 "seek_hole": false, 00:15:03.962 "seek_data": false, 00:15:03.962 "copy": false, 00:15:03.962 "nvme_iov_md": false 00:15:03.962 }, 00:15:03.962 "driver_specific": { 00:15:03.962 "raid": { 00:15:03.962 "uuid": "cf215e54-a9e3-4ce9-9197-3d4f3f032c20", 00:15:03.962 "strip_size_kb": 64, 00:15:03.962 "state": "online", 00:15:03.962 "raid_level": "raid5f", 00:15:03.962 "superblock": true, 00:15:03.962 "num_base_bdevs": 4, 00:15:03.962 "num_base_bdevs_discovered": 4, 00:15:03.962 "num_base_bdevs_operational": 4, 00:15:03.962 "base_bdevs_list": [ 00:15:03.962 { 00:15:03.962 "name": "NewBaseBdev", 00:15:03.962 "uuid": "96da32bf-9504-4923-88f2-61267db13aea", 00:15:03.962 "is_configured": true, 00:15:03.962 "data_offset": 2048, 00:15:03.962 "data_size": 63488 00:15:03.962 }, 00:15:03.962 { 00:15:03.962 "name": "BaseBdev2", 00:15:03.962 "uuid": "a91a1975-222f-4bd5-9ea2-7ad33184546d", 00:15:03.962 "is_configured": true, 00:15:03.962 "data_offset": 2048, 00:15:03.962 "data_size": 63488 00:15:03.962 }, 00:15:03.962 { 00:15:03.962 "name": "BaseBdev3", 00:15:03.962 "uuid": "d0a268f6-25a5-457f-8bcf-a9b560302e8f", 00:15:03.962 "is_configured": true, 00:15:03.962 "data_offset": 2048, 00:15:03.962 "data_size": 63488 00:15:03.962 }, 00:15:03.962 { 00:15:03.962 "name": "BaseBdev4", 00:15:03.962 "uuid": "441e9fb7-5c8f-4c75-beac-3ca4e6d27630", 00:15:03.962 "is_configured": true, 00:15:03.962 "data_offset": 2048, 00:15:03.962 "data_size": 63488 00:15:03.962 } 00:15:03.962 ] 00:15:03.962 } 00:15:03.962 } 00:15:03.962 }' 00:15:03.962 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:04.222 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:04.222 BaseBdev2 00:15:04.222 BaseBdev3 00:15:04.222 BaseBdev4' 00:15:04.222 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:04.222 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:04.222 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:04.222 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.223 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.483 [2024-10-29 11:04:09.726606] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:04.483 [2024-10-29 11:04:09.726689] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:04.483 [2024-10-29 11:04:09.726786] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:04.483 [2024-10-29 11:04:09.727043] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:04.483 [2024-10-29 11:04:09.727055] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name Existed_Raid, state offline 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93999 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@952 -- # '[' -z 93999 ']' 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # kill -0 93999 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@957 -- # uname 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 93999 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 93999' 00:15:04.483 killing process with pid 93999 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@971 -- # kill 93999 00:15:04.483 [2024-10-29 11:04:09.765487] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:04.483 11:04:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@976 -- # wait 93999 00:15:04.483 [2024-10-29 11:04:09.806202] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:04.742 11:04:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:15:04.742 00:15:04.742 real 0m9.601s 00:15:04.742 user 0m16.229s 00:15:04.742 sys 0m2.211s 00:15:04.742 ************************************ 00:15:04.742 END TEST raid5f_state_function_test_sb 00:15:04.742 ************************************ 00:15:04.742 11:04:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:15:04.742 11:04:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.742 11:04:10 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:15:04.742 11:04:10 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:15:04.742 11:04:10 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:15:04.742 11:04:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:04.742 ************************************ 00:15:04.742 START TEST raid5f_superblock_test 00:15:04.742 ************************************ 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1127 -- # raid_superblock_test raid5f 4 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:04.742 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94647 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94647 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@833 -- # '[' -z 94647 ']' 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:15:04.743 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:15:04.743 11:04:10 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.743 [2024-10-29 11:04:10.216028] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:15:04.743 [2024-10-29 11:04:10.216263] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94647 ] 00:15:05.002 [2024-10-29 11:04:10.392090] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:05.002 [2024-10-29 11:04:10.418814] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:05.002 [2024-10-29 11:04:10.463103] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:05.002 [2024-10-29 11:04:10.463235] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@866 -- # return 0 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.574 malloc1 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.574 [2024-10-29 11:04:11.054571] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:05.574 [2024-10-29 11:04:11.054745] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:05.574 [2024-10-29 11:04:11.054790] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:05.574 [2024-10-29 11:04:11.054836] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:05.574 [2024-10-29 11:04:11.056940] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:05.574 [2024-10-29 11:04:11.057040] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:05.574 pt1 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.574 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.835 malloc2 00:15:05.835 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.835 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:05.835 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.835 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.835 [2024-10-29 11:04:11.087240] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:05.835 [2024-10-29 11:04:11.087302] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:05.835 [2024-10-29 11:04:11.087318] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:05.836 [2024-10-29 11:04:11.087330] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:05.836 [2024-10-29 11:04:11.089377] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:05.836 [2024-10-29 11:04:11.089419] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:05.836 pt2 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 malloc3 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 [2024-10-29 11:04:11.115964] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:05.836 [2024-10-29 11:04:11.116101] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:05.836 [2024-10-29 11:04:11.116141] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:05.836 [2024-10-29 11:04:11.116177] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:05.836 [2024-10-29 11:04:11.118248] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:05.836 [2024-10-29 11:04:11.118346] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:05.836 pt3 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 malloc4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 [2024-10-29 11:04:11.163096] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:05.836 [2024-10-29 11:04:11.163321] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:05.836 [2024-10-29 11:04:11.163412] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:05.836 [2024-10-29 11:04:11.163491] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:05.836 [2024-10-29 11:04:11.166974] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:05.836 [2024-10-29 11:04:11.167105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:05.836 pt4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 [2024-10-29 11:04:11.175400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:05.836 [2024-10-29 11:04:11.177694] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:05.836 [2024-10-29 11:04:11.177826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:05.836 [2024-10-29 11:04:11.177941] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:05.836 [2024-10-29 11:04:11.178222] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:05.836 [2024-10-29 11:04:11.178298] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:05.836 [2024-10-29 11:04:11.178711] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:15:05.836 [2024-10-29 11:04:11.179389] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:05.836 [2024-10-29 11:04:11.179453] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:05.836 [2024-10-29 11:04:11.179744] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.836 "name": "raid_bdev1", 00:15:05.836 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:05.836 "strip_size_kb": 64, 00:15:05.836 "state": "online", 00:15:05.836 "raid_level": "raid5f", 00:15:05.836 "superblock": true, 00:15:05.836 "num_base_bdevs": 4, 00:15:05.836 "num_base_bdevs_discovered": 4, 00:15:05.836 "num_base_bdevs_operational": 4, 00:15:05.836 "base_bdevs_list": [ 00:15:05.836 { 00:15:05.836 "name": "pt1", 00:15:05.836 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:05.836 "is_configured": true, 00:15:05.836 "data_offset": 2048, 00:15:05.836 "data_size": 63488 00:15:05.836 }, 00:15:05.836 { 00:15:05.836 "name": "pt2", 00:15:05.836 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:05.836 "is_configured": true, 00:15:05.836 "data_offset": 2048, 00:15:05.836 "data_size": 63488 00:15:05.836 }, 00:15:05.836 { 00:15:05.836 "name": "pt3", 00:15:05.836 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:05.836 "is_configured": true, 00:15:05.836 "data_offset": 2048, 00:15:05.836 "data_size": 63488 00:15:05.836 }, 00:15:05.836 { 00:15:05.836 "name": "pt4", 00:15:05.836 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:05.836 "is_configured": true, 00:15:05.836 "data_offset": 2048, 00:15:05.836 "data_size": 63488 00:15:05.836 } 00:15:05.836 ] 00:15:05.836 }' 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.836 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.407 [2024-10-29 11:04:11.663247] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:06.407 "name": "raid_bdev1", 00:15:06.407 "aliases": [ 00:15:06.407 "1c07554f-8118-4743-ad6f-3f6ffdd85ed4" 00:15:06.407 ], 00:15:06.407 "product_name": "Raid Volume", 00:15:06.407 "block_size": 512, 00:15:06.407 "num_blocks": 190464, 00:15:06.407 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:06.407 "assigned_rate_limits": { 00:15:06.407 "rw_ios_per_sec": 0, 00:15:06.407 "rw_mbytes_per_sec": 0, 00:15:06.407 "r_mbytes_per_sec": 0, 00:15:06.407 "w_mbytes_per_sec": 0 00:15:06.407 }, 00:15:06.407 "claimed": false, 00:15:06.407 "zoned": false, 00:15:06.407 "supported_io_types": { 00:15:06.407 "read": true, 00:15:06.407 "write": true, 00:15:06.407 "unmap": false, 00:15:06.407 "flush": false, 00:15:06.407 "reset": true, 00:15:06.407 "nvme_admin": false, 00:15:06.407 "nvme_io": false, 00:15:06.407 "nvme_io_md": false, 00:15:06.407 "write_zeroes": true, 00:15:06.407 "zcopy": false, 00:15:06.407 "get_zone_info": false, 00:15:06.407 "zone_management": false, 00:15:06.407 "zone_append": false, 00:15:06.407 "compare": false, 00:15:06.407 "compare_and_write": false, 00:15:06.407 "abort": false, 00:15:06.407 "seek_hole": false, 00:15:06.407 "seek_data": false, 00:15:06.407 "copy": false, 00:15:06.407 "nvme_iov_md": false 00:15:06.407 }, 00:15:06.407 "driver_specific": { 00:15:06.407 "raid": { 00:15:06.407 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:06.407 "strip_size_kb": 64, 00:15:06.407 "state": "online", 00:15:06.407 "raid_level": "raid5f", 00:15:06.407 "superblock": true, 00:15:06.407 "num_base_bdevs": 4, 00:15:06.407 "num_base_bdevs_discovered": 4, 00:15:06.407 "num_base_bdevs_operational": 4, 00:15:06.407 "base_bdevs_list": [ 00:15:06.407 { 00:15:06.407 "name": "pt1", 00:15:06.407 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:06.407 "is_configured": true, 00:15:06.407 "data_offset": 2048, 00:15:06.407 "data_size": 63488 00:15:06.407 }, 00:15:06.407 { 00:15:06.407 "name": "pt2", 00:15:06.407 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:06.407 "is_configured": true, 00:15:06.407 "data_offset": 2048, 00:15:06.407 "data_size": 63488 00:15:06.407 }, 00:15:06.407 { 00:15:06.407 "name": "pt3", 00:15:06.407 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:06.407 "is_configured": true, 00:15:06.407 "data_offset": 2048, 00:15:06.407 "data_size": 63488 00:15:06.407 }, 00:15:06.407 { 00:15:06.407 "name": "pt4", 00:15:06.407 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:06.407 "is_configured": true, 00:15:06.407 "data_offset": 2048, 00:15:06.407 "data_size": 63488 00:15:06.407 } 00:15:06.407 ] 00:15:06.407 } 00:15:06.407 } 00:15:06.407 }' 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:06.407 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:06.407 pt2 00:15:06.407 pt3 00:15:06.407 pt4' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.408 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:06.668 11:04:11 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 [2024-10-29 11:04:12.006637] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=1c07554f-8118-4743-ad6f-3f6ffdd85ed4 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 1c07554f-8118-4743-ad6f-3f6ffdd85ed4 ']' 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 [2024-10-29 11:04:12.054406] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:06.668 [2024-10-29 11:04:12.054437] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:06.668 [2024-10-29 11:04:12.054516] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:06.668 [2024-10-29 11:04:12.054601] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:06.668 [2024-10-29 11:04:12.054611] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.668 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:06.669 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.929 [2024-10-29 11:04:12.222131] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:06.929 [2024-10-29 11:04:12.224103] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:06.929 [2024-10-29 11:04:12.224155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:06.929 [2024-10-29 11:04:12.224185] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:15:06.929 [2024-10-29 11:04:12.224231] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:06.929 [2024-10-29 11:04:12.224273] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:06.929 [2024-10-29 11:04:12.224295] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:06.929 [2024-10-29 11:04:12.224314] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:15:06.929 [2024-10-29 11:04:12.224330] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:06.929 [2024-10-29 11:04:12.224350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:15:06.929 request: 00:15:06.929 { 00:15:06.929 "name": "raid_bdev1", 00:15:06.929 "raid_level": "raid5f", 00:15:06.929 "base_bdevs": [ 00:15:06.929 "malloc1", 00:15:06.929 "malloc2", 00:15:06.929 "malloc3", 00:15:06.929 "malloc4" 00:15:06.929 ], 00:15:06.929 "strip_size_kb": 64, 00:15:06.929 "superblock": false, 00:15:06.929 "method": "bdev_raid_create", 00:15:06.929 "req_id": 1 00:15:06.929 } 00:15:06.929 Got JSON-RPC error response 00:15:06.929 response: 00:15:06.929 { 00:15:06.929 "code": -17, 00:15:06.929 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:06.929 } 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.929 [2024-10-29 11:04:12.273992] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:06.929 [2024-10-29 11:04:12.274093] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.929 [2024-10-29 11:04:12.274133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:06.929 [2024-10-29 11:04:12.274163] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.929 [2024-10-29 11:04:12.276290] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.929 [2024-10-29 11:04:12.276393] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:06.929 [2024-10-29 11:04:12.276489] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:06.929 [2024-10-29 11:04:12.276564] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:06.929 pt1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.929 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:06.929 "name": "raid_bdev1", 00:15:06.929 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:06.929 "strip_size_kb": 64, 00:15:06.929 "state": "configuring", 00:15:06.929 "raid_level": "raid5f", 00:15:06.929 "superblock": true, 00:15:06.929 "num_base_bdevs": 4, 00:15:06.929 "num_base_bdevs_discovered": 1, 00:15:06.929 "num_base_bdevs_operational": 4, 00:15:06.929 "base_bdevs_list": [ 00:15:06.929 { 00:15:06.929 "name": "pt1", 00:15:06.930 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:06.930 "is_configured": true, 00:15:06.930 "data_offset": 2048, 00:15:06.930 "data_size": 63488 00:15:06.930 }, 00:15:06.930 { 00:15:06.930 "name": null, 00:15:06.930 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:06.930 "is_configured": false, 00:15:06.930 "data_offset": 2048, 00:15:06.930 "data_size": 63488 00:15:06.930 }, 00:15:06.930 { 00:15:06.930 "name": null, 00:15:06.930 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:06.930 "is_configured": false, 00:15:06.930 "data_offset": 2048, 00:15:06.930 "data_size": 63488 00:15:06.930 }, 00:15:06.930 { 00:15:06.930 "name": null, 00:15:06.930 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:06.930 "is_configured": false, 00:15:06.930 "data_offset": 2048, 00:15:06.930 "data_size": 63488 00:15:06.930 } 00:15:06.930 ] 00:15:06.930 }' 00:15:06.930 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:06.930 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.499 [2024-10-29 11:04:12.721251] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:07.499 [2024-10-29 11:04:12.721305] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.499 [2024-10-29 11:04:12.721324] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:07.499 [2024-10-29 11:04:12.721335] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.499 [2024-10-29 11:04:12.721666] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.499 [2024-10-29 11:04:12.721687] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:07.499 [2024-10-29 11:04:12.721757] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:07.499 [2024-10-29 11:04:12.721774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:07.499 pt2 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.499 [2024-10-29 11:04:12.733241] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.499 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.499 "name": "raid_bdev1", 00:15:07.499 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:07.499 "strip_size_kb": 64, 00:15:07.499 "state": "configuring", 00:15:07.499 "raid_level": "raid5f", 00:15:07.499 "superblock": true, 00:15:07.499 "num_base_bdevs": 4, 00:15:07.499 "num_base_bdevs_discovered": 1, 00:15:07.499 "num_base_bdevs_operational": 4, 00:15:07.499 "base_bdevs_list": [ 00:15:07.499 { 00:15:07.499 "name": "pt1", 00:15:07.499 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:07.499 "is_configured": true, 00:15:07.499 "data_offset": 2048, 00:15:07.499 "data_size": 63488 00:15:07.499 }, 00:15:07.499 { 00:15:07.499 "name": null, 00:15:07.499 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:07.499 "is_configured": false, 00:15:07.499 "data_offset": 0, 00:15:07.499 "data_size": 63488 00:15:07.499 }, 00:15:07.499 { 00:15:07.499 "name": null, 00:15:07.499 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:07.500 "is_configured": false, 00:15:07.500 "data_offset": 2048, 00:15:07.500 "data_size": 63488 00:15:07.500 }, 00:15:07.500 { 00:15:07.500 "name": null, 00:15:07.500 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:07.500 "is_configured": false, 00:15:07.500 "data_offset": 2048, 00:15:07.500 "data_size": 63488 00:15:07.500 } 00:15:07.500 ] 00:15:07.500 }' 00:15:07.500 11:04:12 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.500 11:04:12 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.762 [2024-10-29 11:04:13.212410] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:07.762 [2024-10-29 11:04:13.212521] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.762 [2024-10-29 11:04:13.212557] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:15:07.762 [2024-10-29 11:04:13.212594] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.762 [2024-10-29 11:04:13.213007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.762 [2024-10-29 11:04:13.213076] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:07.762 [2024-10-29 11:04:13.213185] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:07.762 [2024-10-29 11:04:13.213241] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:07.762 pt2 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.762 [2024-10-29 11:04:13.224386] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:07.762 [2024-10-29 11:04:13.224502] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.762 [2024-10-29 11:04:13.224536] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:07.762 [2024-10-29 11:04:13.224571] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.762 [2024-10-29 11:04:13.224914] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.762 [2024-10-29 11:04:13.224983] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:07.762 [2024-10-29 11:04:13.225081] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:07.762 [2024-10-29 11:04:13.225134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:07.762 pt3 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.762 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.762 [2024-10-29 11:04:13.236344] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:07.762 [2024-10-29 11:04:13.236474] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:07.762 [2024-10-29 11:04:13.236510] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:15:07.762 [2024-10-29 11:04:13.236546] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:07.762 [2024-10-29 11:04:13.236880] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:07.763 [2024-10-29 11:04:13.236947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:07.763 [2024-10-29 11:04:13.237045] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:07.763 [2024-10-29 11:04:13.237104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:07.763 [2024-10-29 11:04:13.237257] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:15:07.763 [2024-10-29 11:04:13.237306] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:07.763 [2024-10-29 11:04:13.237564] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:07.763 [2024-10-29 11:04:13.238033] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:15:07.763 [2024-10-29 11:04:13.238045] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:15:07.763 [2024-10-29 11:04:13.238144] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:07.763 pt4 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:07.763 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.028 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.028 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:08.028 "name": "raid_bdev1", 00:15:08.028 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:08.028 "strip_size_kb": 64, 00:15:08.028 "state": "online", 00:15:08.028 "raid_level": "raid5f", 00:15:08.028 "superblock": true, 00:15:08.028 "num_base_bdevs": 4, 00:15:08.028 "num_base_bdevs_discovered": 4, 00:15:08.028 "num_base_bdevs_operational": 4, 00:15:08.028 "base_bdevs_list": [ 00:15:08.028 { 00:15:08.028 "name": "pt1", 00:15:08.028 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:08.028 "is_configured": true, 00:15:08.028 "data_offset": 2048, 00:15:08.028 "data_size": 63488 00:15:08.028 }, 00:15:08.028 { 00:15:08.028 "name": "pt2", 00:15:08.028 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:08.028 "is_configured": true, 00:15:08.028 "data_offset": 2048, 00:15:08.028 "data_size": 63488 00:15:08.028 }, 00:15:08.028 { 00:15:08.028 "name": "pt3", 00:15:08.028 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:08.028 "is_configured": true, 00:15:08.028 "data_offset": 2048, 00:15:08.028 "data_size": 63488 00:15:08.028 }, 00:15:08.028 { 00:15:08.028 "name": "pt4", 00:15:08.028 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:08.028 "is_configured": true, 00:15:08.028 "data_offset": 2048, 00:15:08.028 "data_size": 63488 00:15:08.028 } 00:15:08.028 ] 00:15:08.028 }' 00:15:08.028 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:08.028 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.288 [2024-10-29 11:04:13.715686] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:08.288 "name": "raid_bdev1", 00:15:08.288 "aliases": [ 00:15:08.288 "1c07554f-8118-4743-ad6f-3f6ffdd85ed4" 00:15:08.288 ], 00:15:08.288 "product_name": "Raid Volume", 00:15:08.288 "block_size": 512, 00:15:08.288 "num_blocks": 190464, 00:15:08.288 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:08.288 "assigned_rate_limits": { 00:15:08.288 "rw_ios_per_sec": 0, 00:15:08.288 "rw_mbytes_per_sec": 0, 00:15:08.288 "r_mbytes_per_sec": 0, 00:15:08.288 "w_mbytes_per_sec": 0 00:15:08.288 }, 00:15:08.288 "claimed": false, 00:15:08.288 "zoned": false, 00:15:08.288 "supported_io_types": { 00:15:08.288 "read": true, 00:15:08.288 "write": true, 00:15:08.288 "unmap": false, 00:15:08.288 "flush": false, 00:15:08.288 "reset": true, 00:15:08.288 "nvme_admin": false, 00:15:08.288 "nvme_io": false, 00:15:08.288 "nvme_io_md": false, 00:15:08.288 "write_zeroes": true, 00:15:08.288 "zcopy": false, 00:15:08.288 "get_zone_info": false, 00:15:08.288 "zone_management": false, 00:15:08.288 "zone_append": false, 00:15:08.288 "compare": false, 00:15:08.288 "compare_and_write": false, 00:15:08.288 "abort": false, 00:15:08.288 "seek_hole": false, 00:15:08.288 "seek_data": false, 00:15:08.288 "copy": false, 00:15:08.288 "nvme_iov_md": false 00:15:08.288 }, 00:15:08.288 "driver_specific": { 00:15:08.288 "raid": { 00:15:08.288 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:08.288 "strip_size_kb": 64, 00:15:08.288 "state": "online", 00:15:08.288 "raid_level": "raid5f", 00:15:08.288 "superblock": true, 00:15:08.288 "num_base_bdevs": 4, 00:15:08.288 "num_base_bdevs_discovered": 4, 00:15:08.288 "num_base_bdevs_operational": 4, 00:15:08.288 "base_bdevs_list": [ 00:15:08.288 { 00:15:08.288 "name": "pt1", 00:15:08.288 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:08.288 "is_configured": true, 00:15:08.288 "data_offset": 2048, 00:15:08.288 "data_size": 63488 00:15:08.288 }, 00:15:08.288 { 00:15:08.288 "name": "pt2", 00:15:08.288 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:08.288 "is_configured": true, 00:15:08.288 "data_offset": 2048, 00:15:08.288 "data_size": 63488 00:15:08.288 }, 00:15:08.288 { 00:15:08.288 "name": "pt3", 00:15:08.288 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:08.288 "is_configured": true, 00:15:08.288 "data_offset": 2048, 00:15:08.288 "data_size": 63488 00:15:08.288 }, 00:15:08.288 { 00:15:08.288 "name": "pt4", 00:15:08.288 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:08.288 "is_configured": true, 00:15:08.288 "data_offset": 2048, 00:15:08.288 "data_size": 63488 00:15:08.288 } 00:15:08.288 ] 00:15:08.288 } 00:15:08.288 } 00:15:08.288 }' 00:15:08.288 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:08.289 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:08.289 pt2 00:15:08.289 pt3 00:15:08.289 pt4' 00:15:08.289 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.549 11:04:13 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.549 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.549 [2024-10-29 11:04:14.027172] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 1c07554f-8118-4743-ad6f-3f6ffdd85ed4 '!=' 1c07554f-8118-4743-ad6f-3f6ffdd85ed4 ']' 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.809 [2024-10-29 11:04:14.070922] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:08.809 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.810 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:08.810 "name": "raid_bdev1", 00:15:08.810 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:08.810 "strip_size_kb": 64, 00:15:08.810 "state": "online", 00:15:08.810 "raid_level": "raid5f", 00:15:08.810 "superblock": true, 00:15:08.810 "num_base_bdevs": 4, 00:15:08.810 "num_base_bdevs_discovered": 3, 00:15:08.810 "num_base_bdevs_operational": 3, 00:15:08.810 "base_bdevs_list": [ 00:15:08.810 { 00:15:08.810 "name": null, 00:15:08.810 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:08.810 "is_configured": false, 00:15:08.810 "data_offset": 0, 00:15:08.810 "data_size": 63488 00:15:08.810 }, 00:15:08.810 { 00:15:08.810 "name": "pt2", 00:15:08.810 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:08.810 "is_configured": true, 00:15:08.810 "data_offset": 2048, 00:15:08.810 "data_size": 63488 00:15:08.810 }, 00:15:08.810 { 00:15:08.810 "name": "pt3", 00:15:08.810 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:08.810 "is_configured": true, 00:15:08.810 "data_offset": 2048, 00:15:08.810 "data_size": 63488 00:15:08.810 }, 00:15:08.810 { 00:15:08.810 "name": "pt4", 00:15:08.810 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:08.810 "is_configured": true, 00:15:08.810 "data_offset": 2048, 00:15:08.810 "data_size": 63488 00:15:08.810 } 00:15:08.810 ] 00:15:08.810 }' 00:15:08.810 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:08.810 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.070 [2024-10-29 11:04:14.494239] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:09.070 [2024-10-29 11:04:14.494320] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:09.070 [2024-10-29 11:04:14.494435] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:09.070 [2024-10-29 11:04:14.494560] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:09.070 [2024-10-29 11:04:14.494619] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.070 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.331 [2024-10-29 11:04:14.590070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:09.331 [2024-10-29 11:04:14.590127] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:09.331 [2024-10-29 11:04:14.590143] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:15:09.331 [2024-10-29 11:04:14.590155] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:09.331 [2024-10-29 11:04:14.592287] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:09.331 [2024-10-29 11:04:14.592333] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:09.331 [2024-10-29 11:04:14.592423] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:09.331 [2024-10-29 11:04:14.592459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:09.331 pt2 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.331 "name": "raid_bdev1", 00:15:09.331 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:09.331 "strip_size_kb": 64, 00:15:09.331 "state": "configuring", 00:15:09.331 "raid_level": "raid5f", 00:15:09.331 "superblock": true, 00:15:09.331 "num_base_bdevs": 4, 00:15:09.331 "num_base_bdevs_discovered": 1, 00:15:09.331 "num_base_bdevs_operational": 3, 00:15:09.331 "base_bdevs_list": [ 00:15:09.331 { 00:15:09.331 "name": null, 00:15:09.331 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.331 "is_configured": false, 00:15:09.331 "data_offset": 2048, 00:15:09.331 "data_size": 63488 00:15:09.331 }, 00:15:09.331 { 00:15:09.331 "name": "pt2", 00:15:09.331 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:09.331 "is_configured": true, 00:15:09.331 "data_offset": 2048, 00:15:09.331 "data_size": 63488 00:15:09.331 }, 00:15:09.331 { 00:15:09.331 "name": null, 00:15:09.331 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:09.331 "is_configured": false, 00:15:09.331 "data_offset": 2048, 00:15:09.331 "data_size": 63488 00:15:09.331 }, 00:15:09.331 { 00:15:09.331 "name": null, 00:15:09.331 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:09.331 "is_configured": false, 00:15:09.331 "data_offset": 2048, 00:15:09.331 "data_size": 63488 00:15:09.331 } 00:15:09.331 ] 00:15:09.331 }' 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.331 11:04:14 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.591 [2024-10-29 11:04:15.061257] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:09.591 [2024-10-29 11:04:15.061391] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:09.591 [2024-10-29 11:04:15.061430] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:15:09.591 [2024-10-29 11:04:15.061483] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:09.591 [2024-10-29 11:04:15.061854] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:09.591 [2024-10-29 11:04:15.061924] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:09.591 [2024-10-29 11:04:15.062032] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:09.591 [2024-10-29 11:04:15.062106] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:09.591 pt3 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.591 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.851 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.851 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.851 "name": "raid_bdev1", 00:15:09.851 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:09.851 "strip_size_kb": 64, 00:15:09.851 "state": "configuring", 00:15:09.851 "raid_level": "raid5f", 00:15:09.851 "superblock": true, 00:15:09.851 "num_base_bdevs": 4, 00:15:09.851 "num_base_bdevs_discovered": 2, 00:15:09.851 "num_base_bdevs_operational": 3, 00:15:09.851 "base_bdevs_list": [ 00:15:09.851 { 00:15:09.851 "name": null, 00:15:09.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.851 "is_configured": false, 00:15:09.851 "data_offset": 2048, 00:15:09.851 "data_size": 63488 00:15:09.851 }, 00:15:09.851 { 00:15:09.851 "name": "pt2", 00:15:09.851 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:09.851 "is_configured": true, 00:15:09.851 "data_offset": 2048, 00:15:09.851 "data_size": 63488 00:15:09.851 }, 00:15:09.851 { 00:15:09.851 "name": "pt3", 00:15:09.851 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:09.851 "is_configured": true, 00:15:09.851 "data_offset": 2048, 00:15:09.851 "data_size": 63488 00:15:09.851 }, 00:15:09.851 { 00:15:09.851 "name": null, 00:15:09.851 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:09.851 "is_configured": false, 00:15:09.851 "data_offset": 2048, 00:15:09.851 "data_size": 63488 00:15:09.851 } 00:15:09.851 ] 00:15:09.851 }' 00:15:09.851 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.851 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.111 [2024-10-29 11:04:15.516482] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:10.111 [2024-10-29 11:04:15.516539] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.111 [2024-10-29 11:04:15.516559] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:10.111 [2024-10-29 11:04:15.516572] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.111 [2024-10-29 11:04:15.516884] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.111 [2024-10-29 11:04:15.516905] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:10.111 [2024-10-29 11:04:15.516961] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:10.111 [2024-10-29 11:04:15.516984] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:10.111 [2024-10-29 11:04:15.517074] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:15:10.111 [2024-10-29 11:04:15.517087] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:10.111 [2024-10-29 11:04:15.517308] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:15:10.111 [2024-10-29 11:04:15.517839] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:15:10.111 [2024-10-29 11:04:15.517868] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:15:10.111 [2024-10-29 11:04:15.518133] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:10.111 pt4 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.111 "name": "raid_bdev1", 00:15:10.111 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:10.111 "strip_size_kb": 64, 00:15:10.111 "state": "online", 00:15:10.111 "raid_level": "raid5f", 00:15:10.111 "superblock": true, 00:15:10.111 "num_base_bdevs": 4, 00:15:10.111 "num_base_bdevs_discovered": 3, 00:15:10.111 "num_base_bdevs_operational": 3, 00:15:10.111 "base_bdevs_list": [ 00:15:10.111 { 00:15:10.111 "name": null, 00:15:10.111 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.111 "is_configured": false, 00:15:10.111 "data_offset": 2048, 00:15:10.111 "data_size": 63488 00:15:10.111 }, 00:15:10.111 { 00:15:10.111 "name": "pt2", 00:15:10.111 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:10.111 "is_configured": true, 00:15:10.111 "data_offset": 2048, 00:15:10.111 "data_size": 63488 00:15:10.111 }, 00:15:10.111 { 00:15:10.111 "name": "pt3", 00:15:10.111 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:10.111 "is_configured": true, 00:15:10.111 "data_offset": 2048, 00:15:10.111 "data_size": 63488 00:15:10.111 }, 00:15:10.111 { 00:15:10.111 "name": "pt4", 00:15:10.111 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:10.111 "is_configured": true, 00:15:10.111 "data_offset": 2048, 00:15:10.111 "data_size": 63488 00:15:10.111 } 00:15:10.111 ] 00:15:10.111 }' 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.111 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.681 [2024-10-29 11:04:15.976020] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:10.681 [2024-10-29 11:04:15.976105] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:10.681 [2024-10-29 11:04:15.976180] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:10.681 [2024-10-29 11:04:15.976312] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:10.681 [2024-10-29 11:04:15.976398] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:10.681 11:04:15 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.681 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.681 [2024-10-29 11:04:16.051873] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:10.681 [2024-10-29 11:04:16.052008] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.681 [2024-10-29 11:04:16.052050] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:15:10.681 [2024-10-29 11:04:16.052086] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.681 [2024-10-29 11:04:16.054260] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.681 [2024-10-29 11:04:16.054368] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:10.681 [2024-10-29 11:04:16.054459] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:10.681 [2024-10-29 11:04:16.054539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:10.682 [2024-10-29 11:04:16.054707] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:10.682 [2024-10-29 11:04:16.054773] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:10.682 [2024-10-29 11:04:16.054863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:15:10.682 [2024-10-29 11:04:16.054964] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:10.682 [2024-10-29 11:04:16.055143] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:10.682 pt1 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.682 "name": "raid_bdev1", 00:15:10.682 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:10.682 "strip_size_kb": 64, 00:15:10.682 "state": "configuring", 00:15:10.682 "raid_level": "raid5f", 00:15:10.682 "superblock": true, 00:15:10.682 "num_base_bdevs": 4, 00:15:10.682 "num_base_bdevs_discovered": 2, 00:15:10.682 "num_base_bdevs_operational": 3, 00:15:10.682 "base_bdevs_list": [ 00:15:10.682 { 00:15:10.682 "name": null, 00:15:10.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.682 "is_configured": false, 00:15:10.682 "data_offset": 2048, 00:15:10.682 "data_size": 63488 00:15:10.682 }, 00:15:10.682 { 00:15:10.682 "name": "pt2", 00:15:10.682 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:10.682 "is_configured": true, 00:15:10.682 "data_offset": 2048, 00:15:10.682 "data_size": 63488 00:15:10.682 }, 00:15:10.682 { 00:15:10.682 "name": "pt3", 00:15:10.682 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:10.682 "is_configured": true, 00:15:10.682 "data_offset": 2048, 00:15:10.682 "data_size": 63488 00:15:10.682 }, 00:15:10.682 { 00:15:10.682 "name": null, 00:15:10.682 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:10.682 "is_configured": false, 00:15:10.682 "data_offset": 2048, 00:15:10.682 "data_size": 63488 00:15:10.682 } 00:15:10.682 ] 00:15:10.682 }' 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.682 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.253 [2024-10-29 11:04:16.575020] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:11.253 [2024-10-29 11:04:16.575148] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:11.253 [2024-10-29 11:04:16.575187] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:15:11.253 [2024-10-29 11:04:16.575224] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:11.253 [2024-10-29 11:04:16.575621] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:11.253 [2024-10-29 11:04:16.575694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:11.253 [2024-10-29 11:04:16.575808] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:11.253 [2024-10-29 11:04:16.575865] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:11.253 [2024-10-29 11:04:16.576020] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:15:11.253 [2024-10-29 11:04:16.576070] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:11.253 [2024-10-29 11:04:16.576347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:11.253 [2024-10-29 11:04:16.576944] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:15:11.253 [2024-10-29 11:04:16.577003] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:15:11.253 [2024-10-29 11:04:16.577256] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:11.253 pt4 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.253 "name": "raid_bdev1", 00:15:11.253 "uuid": "1c07554f-8118-4743-ad6f-3f6ffdd85ed4", 00:15:11.253 "strip_size_kb": 64, 00:15:11.253 "state": "online", 00:15:11.253 "raid_level": "raid5f", 00:15:11.253 "superblock": true, 00:15:11.253 "num_base_bdevs": 4, 00:15:11.253 "num_base_bdevs_discovered": 3, 00:15:11.253 "num_base_bdevs_operational": 3, 00:15:11.253 "base_bdevs_list": [ 00:15:11.253 { 00:15:11.253 "name": null, 00:15:11.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:11.253 "is_configured": false, 00:15:11.253 "data_offset": 2048, 00:15:11.253 "data_size": 63488 00:15:11.253 }, 00:15:11.253 { 00:15:11.253 "name": "pt2", 00:15:11.253 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:11.253 "is_configured": true, 00:15:11.253 "data_offset": 2048, 00:15:11.253 "data_size": 63488 00:15:11.253 }, 00:15:11.253 { 00:15:11.253 "name": "pt3", 00:15:11.253 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:11.253 "is_configured": true, 00:15:11.253 "data_offset": 2048, 00:15:11.253 "data_size": 63488 00:15:11.253 }, 00:15:11.253 { 00:15:11.253 "name": "pt4", 00:15:11.253 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:11.253 "is_configured": true, 00:15:11.253 "data_offset": 2048, 00:15:11.253 "data_size": 63488 00:15:11.253 } 00:15:11.253 ] 00:15:11.253 }' 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.253 11:04:16 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.822 [2024-10-29 11:04:17.114345] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 1c07554f-8118-4743-ad6f-3f6ffdd85ed4 '!=' 1c07554f-8118-4743-ad6f-3f6ffdd85ed4 ']' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94647 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@952 -- # '[' -z 94647 ']' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # kill -0 94647 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@957 -- # uname 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 94647 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:15:11.822 killing process with pid 94647 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 94647' 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@971 -- # kill 94647 00:15:11.822 [2024-10-29 11:04:17.174349] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:11.822 [2024-10-29 11:04:17.174431] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:11.822 [2024-10-29 11:04:17.174501] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:11.822 [2024-10-29 11:04:17.174511] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:15:11.822 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@976 -- # wait 94647 00:15:11.822 [2024-10-29 11:04:17.217710] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:12.083 ************************************ 00:15:12.083 END TEST raid5f_superblock_test 00:15:12.083 ************************************ 00:15:12.083 11:04:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:15:12.083 00:15:12.083 real 0m7.324s 00:15:12.083 user 0m12.254s 00:15:12.083 sys 0m1.661s 00:15:12.083 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:15:12.083 11:04:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.083 11:04:17 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:15:12.083 11:04:17 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:15:12.083 11:04:17 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:15:12.083 11:04:17 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:15:12.083 11:04:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:12.083 ************************************ 00:15:12.083 START TEST raid5f_rebuild_test 00:15:12.083 ************************************ 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid5f 4 false false true 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=95121 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 95121 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@833 -- # '[' -z 95121 ']' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # local max_retries=100 00:15:12.083 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # xtrace_disable 00:15:12.083 11:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.343 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:12.343 Zero copy mechanism will not be used. 00:15:12.343 [2024-10-29 11:04:17.632449] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:15:12.343 [2024-10-29 11:04:17.632585] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95121 ] 00:15:12.343 [2024-10-29 11:04:17.805709] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:12.343 [2024-10-29 11:04:17.832614] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:12.601 [2024-10-29 11:04:17.876923] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:12.601 [2024-10-29 11:04:17.876968] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@866 -- # return 0 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.170 BaseBdev1_malloc 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.170 [2024-10-29 11:04:18.464584] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:13.170 [2024-10-29 11:04:18.464665] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.170 [2024-10-29 11:04:18.464690] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:13.170 [2024-10-29 11:04:18.464705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.170 [2024-10-29 11:04:18.466852] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.170 [2024-10-29 11:04:18.466897] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:13.170 BaseBdev1 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.170 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 BaseBdev2_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 [2024-10-29 11:04:18.493317] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:13.171 [2024-10-29 11:04:18.493391] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.171 [2024-10-29 11:04:18.493414] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:13.171 [2024-10-29 11:04:18.493426] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.171 [2024-10-29 11:04:18.495576] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.171 [2024-10-29 11:04:18.495613] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:13.171 BaseBdev2 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 BaseBdev3_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 [2024-10-29 11:04:18.522058] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:13.171 [2024-10-29 11:04:18.522211] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.171 [2024-10-29 11:04:18.522239] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:13.171 [2024-10-29 11:04:18.522250] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.171 [2024-10-29 11:04:18.524405] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.171 [2024-10-29 11:04:18.524445] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:13.171 BaseBdev3 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 BaseBdev4_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 [2024-10-29 11:04:18.568225] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:13.171 [2024-10-29 11:04:18.568335] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.171 [2024-10-29 11:04:18.568428] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:13.171 [2024-10-29 11:04:18.568458] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.171 [2024-10-29 11:04:18.571988] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.171 [2024-10-29 11:04:18.572049] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:13.171 BaseBdev4 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 spare_malloc 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 spare_delay 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 [2024-10-29 11:04:18.609976] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:13.171 [2024-10-29 11:04:18.610135] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.171 [2024-10-29 11:04:18.610179] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:13.171 [2024-10-29 11:04:18.610218] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.171 [2024-10-29 11:04:18.612290] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.171 [2024-10-29 11:04:18.612397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:13.171 spare 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 [2024-10-29 11:04:18.622029] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:13.171 [2024-10-29 11:04:18.623742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:13.171 [2024-10-29 11:04:18.623836] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:13.171 [2024-10-29 11:04:18.623878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:13.171 [2024-10-29 11:04:18.623976] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:13.171 [2024-10-29 11:04:18.624010] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:13.171 [2024-10-29 11:04:18.624267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:13.171 [2024-10-29 11:04:18.624779] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:13.171 [2024-10-29 11:04:18.624807] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:13.171 [2024-10-29 11:04:18.624931] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.171 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.431 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.431 "name": "raid_bdev1", 00:15:13.431 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:13.431 "strip_size_kb": 64, 00:15:13.431 "state": "online", 00:15:13.431 "raid_level": "raid5f", 00:15:13.431 "superblock": false, 00:15:13.431 "num_base_bdevs": 4, 00:15:13.431 "num_base_bdevs_discovered": 4, 00:15:13.431 "num_base_bdevs_operational": 4, 00:15:13.431 "base_bdevs_list": [ 00:15:13.431 { 00:15:13.431 "name": "BaseBdev1", 00:15:13.431 "uuid": "2cb69988-e3f1-566b-9fe4-1d09d5970db5", 00:15:13.431 "is_configured": true, 00:15:13.431 "data_offset": 0, 00:15:13.431 "data_size": 65536 00:15:13.431 }, 00:15:13.431 { 00:15:13.431 "name": "BaseBdev2", 00:15:13.431 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:13.431 "is_configured": true, 00:15:13.431 "data_offset": 0, 00:15:13.431 "data_size": 65536 00:15:13.431 }, 00:15:13.431 { 00:15:13.431 "name": "BaseBdev3", 00:15:13.431 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:13.431 "is_configured": true, 00:15:13.431 "data_offset": 0, 00:15:13.431 "data_size": 65536 00:15:13.431 }, 00:15:13.431 { 00:15:13.431 "name": "BaseBdev4", 00:15:13.431 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:13.431 "is_configured": true, 00:15:13.431 "data_offset": 0, 00:15:13.431 "data_size": 65536 00:15:13.431 } 00:15:13.431 ] 00:15:13.431 }' 00:15:13.431 11:04:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.431 11:04:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.691 [2024-10-29 11:04:19.086075] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:13.691 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:13.957 [2024-10-29 11:04:19.329517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:13.957 /dev/nbd0 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:13.957 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:13.957 1+0 records in 00:15:13.957 1+0 records out 00:15:13.958 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000535333 s, 7.7 MB/s 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:13.958 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:15:14.558 512+0 records in 00:15:14.558 512+0 records out 00:15:14.558 100663296 bytes (101 MB, 96 MiB) copied, 0.513076 s, 196 MB/s 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:14.558 11:04:19 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:14.817 [2024-10-29 11:04:20.141481] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.817 [2024-10-29 11:04:20.156143] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.817 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.817 "name": "raid_bdev1", 00:15:14.817 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:14.817 "strip_size_kb": 64, 00:15:14.817 "state": "online", 00:15:14.817 "raid_level": "raid5f", 00:15:14.817 "superblock": false, 00:15:14.817 "num_base_bdevs": 4, 00:15:14.817 "num_base_bdevs_discovered": 3, 00:15:14.817 "num_base_bdevs_operational": 3, 00:15:14.817 "base_bdevs_list": [ 00:15:14.817 { 00:15:14.817 "name": null, 00:15:14.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:14.817 "is_configured": false, 00:15:14.817 "data_offset": 0, 00:15:14.817 "data_size": 65536 00:15:14.817 }, 00:15:14.817 { 00:15:14.817 "name": "BaseBdev2", 00:15:14.817 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:14.817 "is_configured": true, 00:15:14.817 "data_offset": 0, 00:15:14.817 "data_size": 65536 00:15:14.817 }, 00:15:14.817 { 00:15:14.817 "name": "BaseBdev3", 00:15:14.817 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:14.817 "is_configured": true, 00:15:14.817 "data_offset": 0, 00:15:14.817 "data_size": 65536 00:15:14.818 }, 00:15:14.818 { 00:15:14.818 "name": "BaseBdev4", 00:15:14.818 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:14.818 "is_configured": true, 00:15:14.818 "data_offset": 0, 00:15:14.818 "data_size": 65536 00:15:14.818 } 00:15:14.818 ] 00:15:14.818 }' 00:15:14.818 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.818 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.387 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:15.387 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.387 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.387 [2024-10-29 11:04:20.591789] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:15.387 [2024-10-29 11:04:20.596278] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b5b0 00:15:15.387 11:04:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.387 11:04:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:15.387 [2024-10-29 11:04:20.598548] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:16.325 "name": "raid_bdev1", 00:15:16.325 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:16.325 "strip_size_kb": 64, 00:15:16.325 "state": "online", 00:15:16.325 "raid_level": "raid5f", 00:15:16.325 "superblock": false, 00:15:16.325 "num_base_bdevs": 4, 00:15:16.325 "num_base_bdevs_discovered": 4, 00:15:16.325 "num_base_bdevs_operational": 4, 00:15:16.325 "process": { 00:15:16.325 "type": "rebuild", 00:15:16.325 "target": "spare", 00:15:16.325 "progress": { 00:15:16.325 "blocks": 19200, 00:15:16.325 "percent": 9 00:15:16.325 } 00:15:16.325 }, 00:15:16.325 "base_bdevs_list": [ 00:15:16.325 { 00:15:16.325 "name": "spare", 00:15:16.325 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:16.325 "is_configured": true, 00:15:16.325 "data_offset": 0, 00:15:16.325 "data_size": 65536 00:15:16.325 }, 00:15:16.325 { 00:15:16.325 "name": "BaseBdev2", 00:15:16.325 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:16.325 "is_configured": true, 00:15:16.325 "data_offset": 0, 00:15:16.325 "data_size": 65536 00:15:16.325 }, 00:15:16.325 { 00:15:16.325 "name": "BaseBdev3", 00:15:16.325 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:16.325 "is_configured": true, 00:15:16.325 "data_offset": 0, 00:15:16.325 "data_size": 65536 00:15:16.325 }, 00:15:16.325 { 00:15:16.325 "name": "BaseBdev4", 00:15:16.325 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:16.325 "is_configured": true, 00:15:16.325 "data_offset": 0, 00:15:16.325 "data_size": 65536 00:15:16.325 } 00:15:16.325 ] 00:15:16.325 }' 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:16.325 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.326 [2024-10-29 11:04:21.739162] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:16.326 [2024-10-29 11:04:21.804267] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:16.326 [2024-10-29 11:04:21.804325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:16.326 [2024-10-29 11:04:21.804345] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:16.326 [2024-10-29 11:04:21.804353] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.326 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.585 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.585 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.585 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:16.585 "name": "raid_bdev1", 00:15:16.585 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:16.585 "strip_size_kb": 64, 00:15:16.585 "state": "online", 00:15:16.585 "raid_level": "raid5f", 00:15:16.585 "superblock": false, 00:15:16.585 "num_base_bdevs": 4, 00:15:16.585 "num_base_bdevs_discovered": 3, 00:15:16.585 "num_base_bdevs_operational": 3, 00:15:16.585 "base_bdevs_list": [ 00:15:16.585 { 00:15:16.585 "name": null, 00:15:16.585 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.585 "is_configured": false, 00:15:16.585 "data_offset": 0, 00:15:16.585 "data_size": 65536 00:15:16.585 }, 00:15:16.585 { 00:15:16.585 "name": "BaseBdev2", 00:15:16.585 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:16.585 "is_configured": true, 00:15:16.585 "data_offset": 0, 00:15:16.585 "data_size": 65536 00:15:16.585 }, 00:15:16.585 { 00:15:16.585 "name": "BaseBdev3", 00:15:16.585 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:16.585 "is_configured": true, 00:15:16.585 "data_offset": 0, 00:15:16.585 "data_size": 65536 00:15:16.585 }, 00:15:16.585 { 00:15:16.586 "name": "BaseBdev4", 00:15:16.586 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:16.586 "is_configured": true, 00:15:16.586 "data_offset": 0, 00:15:16.586 "data_size": 65536 00:15:16.586 } 00:15:16.586 ] 00:15:16.586 }' 00:15:16.586 11:04:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:16.586 11:04:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.845 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:17.105 "name": "raid_bdev1", 00:15:17.105 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:17.105 "strip_size_kb": 64, 00:15:17.105 "state": "online", 00:15:17.105 "raid_level": "raid5f", 00:15:17.105 "superblock": false, 00:15:17.105 "num_base_bdevs": 4, 00:15:17.105 "num_base_bdevs_discovered": 3, 00:15:17.105 "num_base_bdevs_operational": 3, 00:15:17.105 "base_bdevs_list": [ 00:15:17.105 { 00:15:17.105 "name": null, 00:15:17.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:17.105 "is_configured": false, 00:15:17.105 "data_offset": 0, 00:15:17.105 "data_size": 65536 00:15:17.105 }, 00:15:17.105 { 00:15:17.105 "name": "BaseBdev2", 00:15:17.105 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:17.105 "is_configured": true, 00:15:17.105 "data_offset": 0, 00:15:17.105 "data_size": 65536 00:15:17.105 }, 00:15:17.105 { 00:15:17.105 "name": "BaseBdev3", 00:15:17.105 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:17.105 "is_configured": true, 00:15:17.105 "data_offset": 0, 00:15:17.105 "data_size": 65536 00:15:17.105 }, 00:15:17.105 { 00:15:17.105 "name": "BaseBdev4", 00:15:17.105 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:17.105 "is_configured": true, 00:15:17.105 "data_offset": 0, 00:15:17.105 "data_size": 65536 00:15:17.105 } 00:15:17.105 ] 00:15:17.105 }' 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.105 [2024-10-29 11:04:22.429243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:17.105 [2024-10-29 11:04:22.432770] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b680 00:15:17.105 [2024-10-29 11:04:22.434981] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.105 11:04:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:18.045 "name": "raid_bdev1", 00:15:18.045 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:18.045 "strip_size_kb": 64, 00:15:18.045 "state": "online", 00:15:18.045 "raid_level": "raid5f", 00:15:18.045 "superblock": false, 00:15:18.045 "num_base_bdevs": 4, 00:15:18.045 "num_base_bdevs_discovered": 4, 00:15:18.045 "num_base_bdevs_operational": 4, 00:15:18.045 "process": { 00:15:18.045 "type": "rebuild", 00:15:18.045 "target": "spare", 00:15:18.045 "progress": { 00:15:18.045 "blocks": 19200, 00:15:18.045 "percent": 9 00:15:18.045 } 00:15:18.045 }, 00:15:18.045 "base_bdevs_list": [ 00:15:18.045 { 00:15:18.045 "name": "spare", 00:15:18.045 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:18.045 "is_configured": true, 00:15:18.045 "data_offset": 0, 00:15:18.045 "data_size": 65536 00:15:18.045 }, 00:15:18.045 { 00:15:18.045 "name": "BaseBdev2", 00:15:18.045 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:18.045 "is_configured": true, 00:15:18.045 "data_offset": 0, 00:15:18.045 "data_size": 65536 00:15:18.045 }, 00:15:18.045 { 00:15:18.045 "name": "BaseBdev3", 00:15:18.045 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:18.045 "is_configured": true, 00:15:18.045 "data_offset": 0, 00:15:18.045 "data_size": 65536 00:15:18.045 }, 00:15:18.045 { 00:15:18.045 "name": "BaseBdev4", 00:15:18.045 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:18.045 "is_configured": true, 00:15:18.045 "data_offset": 0, 00:15:18.045 "data_size": 65536 00:15:18.045 } 00:15:18.045 ] 00:15:18.045 }' 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:18.045 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=508 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:18.306 "name": "raid_bdev1", 00:15:18.306 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:18.306 "strip_size_kb": 64, 00:15:18.306 "state": "online", 00:15:18.306 "raid_level": "raid5f", 00:15:18.306 "superblock": false, 00:15:18.306 "num_base_bdevs": 4, 00:15:18.306 "num_base_bdevs_discovered": 4, 00:15:18.306 "num_base_bdevs_operational": 4, 00:15:18.306 "process": { 00:15:18.306 "type": "rebuild", 00:15:18.306 "target": "spare", 00:15:18.306 "progress": { 00:15:18.306 "blocks": 21120, 00:15:18.306 "percent": 10 00:15:18.306 } 00:15:18.306 }, 00:15:18.306 "base_bdevs_list": [ 00:15:18.306 { 00:15:18.306 "name": "spare", 00:15:18.306 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:18.306 "is_configured": true, 00:15:18.306 "data_offset": 0, 00:15:18.306 "data_size": 65536 00:15:18.306 }, 00:15:18.306 { 00:15:18.306 "name": "BaseBdev2", 00:15:18.306 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:18.306 "is_configured": true, 00:15:18.306 "data_offset": 0, 00:15:18.306 "data_size": 65536 00:15:18.306 }, 00:15:18.306 { 00:15:18.306 "name": "BaseBdev3", 00:15:18.306 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:18.306 "is_configured": true, 00:15:18.306 "data_offset": 0, 00:15:18.306 "data_size": 65536 00:15:18.306 }, 00:15:18.306 { 00:15:18.306 "name": "BaseBdev4", 00:15:18.306 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:18.306 "is_configured": true, 00:15:18.306 "data_offset": 0, 00:15:18.306 "data_size": 65536 00:15:18.306 } 00:15:18.306 ] 00:15:18.306 }' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:18.306 11:04:23 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.247 11:04:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:19.507 "name": "raid_bdev1", 00:15:19.507 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:19.507 "strip_size_kb": 64, 00:15:19.507 "state": "online", 00:15:19.507 "raid_level": "raid5f", 00:15:19.507 "superblock": false, 00:15:19.507 "num_base_bdevs": 4, 00:15:19.507 "num_base_bdevs_discovered": 4, 00:15:19.507 "num_base_bdevs_operational": 4, 00:15:19.507 "process": { 00:15:19.507 "type": "rebuild", 00:15:19.507 "target": "spare", 00:15:19.507 "progress": { 00:15:19.507 "blocks": 42240, 00:15:19.507 "percent": 21 00:15:19.507 } 00:15:19.507 }, 00:15:19.507 "base_bdevs_list": [ 00:15:19.507 { 00:15:19.507 "name": "spare", 00:15:19.507 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:19.507 "is_configured": true, 00:15:19.507 "data_offset": 0, 00:15:19.507 "data_size": 65536 00:15:19.507 }, 00:15:19.507 { 00:15:19.507 "name": "BaseBdev2", 00:15:19.507 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:19.507 "is_configured": true, 00:15:19.507 "data_offset": 0, 00:15:19.507 "data_size": 65536 00:15:19.507 }, 00:15:19.507 { 00:15:19.507 "name": "BaseBdev3", 00:15:19.507 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:19.507 "is_configured": true, 00:15:19.507 "data_offset": 0, 00:15:19.507 "data_size": 65536 00:15:19.507 }, 00:15:19.507 { 00:15:19.507 "name": "BaseBdev4", 00:15:19.507 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:19.507 "is_configured": true, 00:15:19.507 "data_offset": 0, 00:15:19.507 "data_size": 65536 00:15:19.507 } 00:15:19.507 ] 00:15:19.507 }' 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:19.507 11:04:24 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:20.448 "name": "raid_bdev1", 00:15:20.448 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:20.448 "strip_size_kb": 64, 00:15:20.448 "state": "online", 00:15:20.448 "raid_level": "raid5f", 00:15:20.448 "superblock": false, 00:15:20.448 "num_base_bdevs": 4, 00:15:20.448 "num_base_bdevs_discovered": 4, 00:15:20.448 "num_base_bdevs_operational": 4, 00:15:20.448 "process": { 00:15:20.448 "type": "rebuild", 00:15:20.448 "target": "spare", 00:15:20.448 "progress": { 00:15:20.448 "blocks": 65280, 00:15:20.448 "percent": 33 00:15:20.448 } 00:15:20.448 }, 00:15:20.448 "base_bdevs_list": [ 00:15:20.448 { 00:15:20.448 "name": "spare", 00:15:20.448 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:20.448 "is_configured": true, 00:15:20.448 "data_offset": 0, 00:15:20.448 "data_size": 65536 00:15:20.448 }, 00:15:20.448 { 00:15:20.448 "name": "BaseBdev2", 00:15:20.448 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:20.448 "is_configured": true, 00:15:20.448 "data_offset": 0, 00:15:20.448 "data_size": 65536 00:15:20.448 }, 00:15:20.448 { 00:15:20.448 "name": "BaseBdev3", 00:15:20.448 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:20.448 "is_configured": true, 00:15:20.448 "data_offset": 0, 00:15:20.448 "data_size": 65536 00:15:20.448 }, 00:15:20.448 { 00:15:20.448 "name": "BaseBdev4", 00:15:20.448 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:20.448 "is_configured": true, 00:15:20.448 "data_offset": 0, 00:15:20.448 "data_size": 65536 00:15:20.448 } 00:15:20.448 ] 00:15:20.448 }' 00:15:20.448 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:20.708 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:20.708 11:04:25 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:20.708 11:04:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:20.708 11:04:26 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.647 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.647 "name": "raid_bdev1", 00:15:21.647 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:21.647 "strip_size_kb": 64, 00:15:21.647 "state": "online", 00:15:21.647 "raid_level": "raid5f", 00:15:21.647 "superblock": false, 00:15:21.647 "num_base_bdevs": 4, 00:15:21.647 "num_base_bdevs_discovered": 4, 00:15:21.647 "num_base_bdevs_operational": 4, 00:15:21.647 "process": { 00:15:21.647 "type": "rebuild", 00:15:21.647 "target": "spare", 00:15:21.648 "progress": { 00:15:21.648 "blocks": 86400, 00:15:21.648 "percent": 43 00:15:21.648 } 00:15:21.648 }, 00:15:21.648 "base_bdevs_list": [ 00:15:21.648 { 00:15:21.648 "name": "spare", 00:15:21.648 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:21.648 "is_configured": true, 00:15:21.648 "data_offset": 0, 00:15:21.648 "data_size": 65536 00:15:21.648 }, 00:15:21.648 { 00:15:21.648 "name": "BaseBdev2", 00:15:21.648 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:21.648 "is_configured": true, 00:15:21.648 "data_offset": 0, 00:15:21.648 "data_size": 65536 00:15:21.648 }, 00:15:21.648 { 00:15:21.648 "name": "BaseBdev3", 00:15:21.648 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:21.648 "is_configured": true, 00:15:21.648 "data_offset": 0, 00:15:21.648 "data_size": 65536 00:15:21.648 }, 00:15:21.648 { 00:15:21.648 "name": "BaseBdev4", 00:15:21.648 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:21.648 "is_configured": true, 00:15:21.648 "data_offset": 0, 00:15:21.648 "data_size": 65536 00:15:21.648 } 00:15:21.648 ] 00:15:21.648 }' 00:15:21.648 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.648 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:21.648 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.907 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:21.907 11:04:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.847 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.847 "name": "raid_bdev1", 00:15:22.847 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:22.847 "strip_size_kb": 64, 00:15:22.847 "state": "online", 00:15:22.847 "raid_level": "raid5f", 00:15:22.848 "superblock": false, 00:15:22.848 "num_base_bdevs": 4, 00:15:22.848 "num_base_bdevs_discovered": 4, 00:15:22.848 "num_base_bdevs_operational": 4, 00:15:22.848 "process": { 00:15:22.848 "type": "rebuild", 00:15:22.848 "target": "spare", 00:15:22.848 "progress": { 00:15:22.848 "blocks": 109440, 00:15:22.848 "percent": 55 00:15:22.848 } 00:15:22.848 }, 00:15:22.848 "base_bdevs_list": [ 00:15:22.848 { 00:15:22.848 "name": "spare", 00:15:22.848 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:22.848 "is_configured": true, 00:15:22.848 "data_offset": 0, 00:15:22.848 "data_size": 65536 00:15:22.848 }, 00:15:22.848 { 00:15:22.848 "name": "BaseBdev2", 00:15:22.848 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:22.848 "is_configured": true, 00:15:22.848 "data_offset": 0, 00:15:22.848 "data_size": 65536 00:15:22.848 }, 00:15:22.848 { 00:15:22.848 "name": "BaseBdev3", 00:15:22.848 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:22.848 "is_configured": true, 00:15:22.848 "data_offset": 0, 00:15:22.848 "data_size": 65536 00:15:22.848 }, 00:15:22.848 { 00:15:22.848 "name": "BaseBdev4", 00:15:22.848 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:22.848 "is_configured": true, 00:15:22.848 "data_offset": 0, 00:15:22.848 "data_size": 65536 00:15:22.848 } 00:15:22.848 ] 00:15:22.848 }' 00:15:22.848 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.848 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:22.848 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.848 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:22.848 11:04:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:24.234 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:24.234 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.234 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.235 "name": "raid_bdev1", 00:15:24.235 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:24.235 "strip_size_kb": 64, 00:15:24.235 "state": "online", 00:15:24.235 "raid_level": "raid5f", 00:15:24.235 "superblock": false, 00:15:24.235 "num_base_bdevs": 4, 00:15:24.235 "num_base_bdevs_discovered": 4, 00:15:24.235 "num_base_bdevs_operational": 4, 00:15:24.235 "process": { 00:15:24.235 "type": "rebuild", 00:15:24.235 "target": "spare", 00:15:24.235 "progress": { 00:15:24.235 "blocks": 130560, 00:15:24.235 "percent": 66 00:15:24.235 } 00:15:24.235 }, 00:15:24.235 "base_bdevs_list": [ 00:15:24.235 { 00:15:24.235 "name": "spare", 00:15:24.235 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:24.235 "is_configured": true, 00:15:24.235 "data_offset": 0, 00:15:24.235 "data_size": 65536 00:15:24.235 }, 00:15:24.235 { 00:15:24.235 "name": "BaseBdev2", 00:15:24.235 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:24.235 "is_configured": true, 00:15:24.235 "data_offset": 0, 00:15:24.235 "data_size": 65536 00:15:24.235 }, 00:15:24.235 { 00:15:24.235 "name": "BaseBdev3", 00:15:24.235 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:24.235 "is_configured": true, 00:15:24.235 "data_offset": 0, 00:15:24.235 "data_size": 65536 00:15:24.235 }, 00:15:24.235 { 00:15:24.235 "name": "BaseBdev4", 00:15:24.235 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:24.235 "is_configured": true, 00:15:24.235 "data_offset": 0, 00:15:24.235 "data_size": 65536 00:15:24.235 } 00:15:24.235 ] 00:15:24.235 }' 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.235 11:04:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.179 "name": "raid_bdev1", 00:15:25.179 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:25.179 "strip_size_kb": 64, 00:15:25.179 "state": "online", 00:15:25.179 "raid_level": "raid5f", 00:15:25.179 "superblock": false, 00:15:25.179 "num_base_bdevs": 4, 00:15:25.179 "num_base_bdevs_discovered": 4, 00:15:25.179 "num_base_bdevs_operational": 4, 00:15:25.179 "process": { 00:15:25.179 "type": "rebuild", 00:15:25.179 "target": "spare", 00:15:25.179 "progress": { 00:15:25.179 "blocks": 153600, 00:15:25.179 "percent": 78 00:15:25.179 } 00:15:25.179 }, 00:15:25.179 "base_bdevs_list": [ 00:15:25.179 { 00:15:25.179 "name": "spare", 00:15:25.179 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:25.179 "is_configured": true, 00:15:25.179 "data_offset": 0, 00:15:25.179 "data_size": 65536 00:15:25.179 }, 00:15:25.179 { 00:15:25.179 "name": "BaseBdev2", 00:15:25.179 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:25.179 "is_configured": true, 00:15:25.179 "data_offset": 0, 00:15:25.179 "data_size": 65536 00:15:25.179 }, 00:15:25.179 { 00:15:25.179 "name": "BaseBdev3", 00:15:25.179 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:25.179 "is_configured": true, 00:15:25.179 "data_offset": 0, 00:15:25.179 "data_size": 65536 00:15:25.179 }, 00:15:25.179 { 00:15:25.179 "name": "BaseBdev4", 00:15:25.179 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:25.179 "is_configured": true, 00:15:25.179 "data_offset": 0, 00:15:25.179 "data_size": 65536 00:15:25.179 } 00:15:25.179 ] 00:15:25.179 }' 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.179 11:04:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:26.562 "name": "raid_bdev1", 00:15:26.562 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:26.562 "strip_size_kb": 64, 00:15:26.562 "state": "online", 00:15:26.562 "raid_level": "raid5f", 00:15:26.562 "superblock": false, 00:15:26.562 "num_base_bdevs": 4, 00:15:26.562 "num_base_bdevs_discovered": 4, 00:15:26.562 "num_base_bdevs_operational": 4, 00:15:26.562 "process": { 00:15:26.562 "type": "rebuild", 00:15:26.562 "target": "spare", 00:15:26.562 "progress": { 00:15:26.562 "blocks": 174720, 00:15:26.562 "percent": 88 00:15:26.562 } 00:15:26.562 }, 00:15:26.562 "base_bdevs_list": [ 00:15:26.562 { 00:15:26.562 "name": "spare", 00:15:26.562 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:26.562 "is_configured": true, 00:15:26.562 "data_offset": 0, 00:15:26.562 "data_size": 65536 00:15:26.562 }, 00:15:26.562 { 00:15:26.562 "name": "BaseBdev2", 00:15:26.562 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:26.562 "is_configured": true, 00:15:26.562 "data_offset": 0, 00:15:26.562 "data_size": 65536 00:15:26.562 }, 00:15:26.562 { 00:15:26.562 "name": "BaseBdev3", 00:15:26.562 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:26.562 "is_configured": true, 00:15:26.562 "data_offset": 0, 00:15:26.562 "data_size": 65536 00:15:26.562 }, 00:15:26.562 { 00:15:26.562 "name": "BaseBdev4", 00:15:26.562 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:26.562 "is_configured": true, 00:15:26.562 "data_offset": 0, 00:15:26.562 "data_size": 65536 00:15:26.562 } 00:15:26.562 ] 00:15:26.562 }' 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:26.562 11:04:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:27.502 [2024-10-29 11:04:32.776036] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:27.502 [2024-10-29 11:04:32.776177] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:27.502 [2024-10-29 11:04:32.776248] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.502 "name": "raid_bdev1", 00:15:27.502 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:27.502 "strip_size_kb": 64, 00:15:27.502 "state": "online", 00:15:27.502 "raid_level": "raid5f", 00:15:27.502 "superblock": false, 00:15:27.502 "num_base_bdevs": 4, 00:15:27.502 "num_base_bdevs_discovered": 4, 00:15:27.502 "num_base_bdevs_operational": 4, 00:15:27.502 "process": { 00:15:27.502 "type": "rebuild", 00:15:27.502 "target": "spare", 00:15:27.502 "progress": { 00:15:27.502 "blocks": 195840, 00:15:27.502 "percent": 99 00:15:27.502 } 00:15:27.502 }, 00:15:27.502 "base_bdevs_list": [ 00:15:27.502 { 00:15:27.502 "name": "spare", 00:15:27.502 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:27.502 "is_configured": true, 00:15:27.502 "data_offset": 0, 00:15:27.502 "data_size": 65536 00:15:27.502 }, 00:15:27.502 { 00:15:27.502 "name": "BaseBdev2", 00:15:27.502 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:27.502 "is_configured": true, 00:15:27.502 "data_offset": 0, 00:15:27.502 "data_size": 65536 00:15:27.502 }, 00:15:27.502 { 00:15:27.502 "name": "BaseBdev3", 00:15:27.502 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:27.502 "is_configured": true, 00:15:27.502 "data_offset": 0, 00:15:27.502 "data_size": 65536 00:15:27.502 }, 00:15:27.502 { 00:15:27.502 "name": "BaseBdev4", 00:15:27.502 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:27.502 "is_configured": true, 00:15:27.502 "data_offset": 0, 00:15:27.502 "data_size": 65536 00:15:27.502 } 00:15:27.502 ] 00:15:27.502 }' 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:27.502 11:04:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.442 11:04:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:28.703 11:04:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.703 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:28.703 "name": "raid_bdev1", 00:15:28.703 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:28.703 "strip_size_kb": 64, 00:15:28.703 "state": "online", 00:15:28.703 "raid_level": "raid5f", 00:15:28.703 "superblock": false, 00:15:28.703 "num_base_bdevs": 4, 00:15:28.703 "num_base_bdevs_discovered": 4, 00:15:28.703 "num_base_bdevs_operational": 4, 00:15:28.703 "base_bdevs_list": [ 00:15:28.703 { 00:15:28.703 "name": "spare", 00:15:28.703 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev2", 00:15:28.703 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev3", 00:15:28.703 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev4", 00:15:28.703 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 } 00:15:28.703 ] 00:15:28.703 }' 00:15:28.703 11:04:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:28.703 "name": "raid_bdev1", 00:15:28.703 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:28.703 "strip_size_kb": 64, 00:15:28.703 "state": "online", 00:15:28.703 "raid_level": "raid5f", 00:15:28.703 "superblock": false, 00:15:28.703 "num_base_bdevs": 4, 00:15:28.703 "num_base_bdevs_discovered": 4, 00:15:28.703 "num_base_bdevs_operational": 4, 00:15:28.703 "base_bdevs_list": [ 00:15:28.703 { 00:15:28.703 "name": "spare", 00:15:28.703 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev2", 00:15:28.703 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev3", 00:15:28.703 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 }, 00:15:28.703 { 00:15:28.703 "name": "BaseBdev4", 00:15:28.703 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:28.703 "is_configured": true, 00:15:28.703 "data_offset": 0, 00:15:28.703 "data_size": 65536 00:15:28.703 } 00:15:28.703 ] 00:15:28.703 }' 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:28.703 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:28.963 "name": "raid_bdev1", 00:15:28.963 "uuid": "65ad9f5d-0686-41f6-9335-f34b180401f4", 00:15:28.963 "strip_size_kb": 64, 00:15:28.963 "state": "online", 00:15:28.963 "raid_level": "raid5f", 00:15:28.963 "superblock": false, 00:15:28.963 "num_base_bdevs": 4, 00:15:28.963 "num_base_bdevs_discovered": 4, 00:15:28.963 "num_base_bdevs_operational": 4, 00:15:28.963 "base_bdevs_list": [ 00:15:28.963 { 00:15:28.963 "name": "spare", 00:15:28.963 "uuid": "cd902561-e12c-5b9b-9d50-4497f80c032b", 00:15:28.963 "is_configured": true, 00:15:28.963 "data_offset": 0, 00:15:28.963 "data_size": 65536 00:15:28.963 }, 00:15:28.963 { 00:15:28.963 "name": "BaseBdev2", 00:15:28.963 "uuid": "7a3ea945-89bb-5a09-9852-a16c1cba462d", 00:15:28.963 "is_configured": true, 00:15:28.963 "data_offset": 0, 00:15:28.963 "data_size": 65536 00:15:28.963 }, 00:15:28.963 { 00:15:28.963 "name": "BaseBdev3", 00:15:28.963 "uuid": "df32581a-8042-5d8c-9504-effe452489ef", 00:15:28.963 "is_configured": true, 00:15:28.963 "data_offset": 0, 00:15:28.963 "data_size": 65536 00:15:28.963 }, 00:15:28.963 { 00:15:28.963 "name": "BaseBdev4", 00:15:28.963 "uuid": "c55e7e27-85b5-5854-ae90-25981f55e853", 00:15:28.963 "is_configured": true, 00:15:28.963 "data_offset": 0, 00:15:28.963 "data_size": 65536 00:15:28.963 } 00:15:28.963 ] 00:15:28.963 }' 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:28.963 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:29.224 [2024-10-29 11:04:34.678608] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:29.224 [2024-10-29 11:04:34.678683] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:29.224 [2024-10-29 11:04:34.678789] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:29.224 [2024-10-29 11:04:34.678964] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:29.224 [2024-10-29 11:04:34.679021] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:29.224 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:29.484 /dev/nbd0 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:29.484 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:29.485 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:29.485 1+0 records in 00:15:29.485 1+0 records out 00:15:29.485 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000605105 s, 6.8 MB/s 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:29.745 11:04:34 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:29.745 /dev/nbd1 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # local i 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # break 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:29.745 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:29.745 1+0 records in 00:15:29.745 1+0 records out 00:15:29.746 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000369858 s, 11.1 MB/s 00:15:29.746 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:29.746 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # size=4096 00:15:29.746 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # return 0 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:30.006 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 95121 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@952 -- # '[' -z 95121 ']' 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # kill -0 95121 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@957 -- # uname 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:15:30.267 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 95121 00:15:30.527 killing process with pid 95121 00:15:30.527 Received shutdown signal, test time was about 60.000000 seconds 00:15:30.527 00:15:30.527 Latency(us) 00:15:30.527 [2024-10-29T11:04:36.025Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:30.527 [2024-10-29T11:04:36.025Z] =================================================================================================================== 00:15:30.527 [2024-10-29T11:04:36.025Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:30.527 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:15:30.527 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:15:30.527 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@970 -- # echo 'killing process with pid 95121' 00:15:30.527 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@971 -- # kill 95121 00:15:30.527 [2024-10-29 11:04:35.775319] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:30.527 11:04:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@976 -- # wait 95121 00:15:30.527 [2024-10-29 11:04:35.825892] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:30.788 00:15:30.788 real 0m18.500s 00:15:30.788 user 0m22.236s 00:15:30.788 sys 0m2.463s 00:15:30.788 ************************************ 00:15:30.788 END TEST raid5f_rebuild_test 00:15:30.788 ************************************ 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1128 -- # xtrace_disable 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:30.788 11:04:36 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:15:30.788 11:04:36 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:15:30.788 11:04:36 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:15:30.788 11:04:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:30.788 ************************************ 00:15:30.788 START TEST raid5f_rebuild_test_sb 00:15:30.788 ************************************ 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid5f 4 true false true 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95621 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95621 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@833 -- # '[' -z 95621 ']' 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # local max_retries=100 00:15:30.788 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:30.789 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:30.789 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # xtrace_disable 00:15:30.789 11:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.789 [2024-10-29 11:04:36.222908] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:15:30.789 [2024-10-29 11:04:36.223116] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95621 ] 00:15:30.789 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:30.789 Zero copy mechanism will not be used. 00:15:31.049 [2024-10-29 11:04:36.395282] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:31.049 [2024-10-29 11:04:36.421920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:31.049 [2024-10-29 11:04:36.467283] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:31.049 [2024-10-29 11:04:36.467422] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@866 -- # return 0 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.621 BaseBdev1_malloc 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.621 [2024-10-29 11:04:37.075175] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:31.621 [2024-10-29 11:04:37.075302] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.621 [2024-10-29 11:04:37.075368] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:31.621 [2024-10-29 11:04:37.075457] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.621 [2024-10-29 11:04:37.077762] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.621 [2024-10-29 11:04:37.077853] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:31.621 BaseBdev1 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.621 BaseBdev2_malloc 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.621 [2024-10-29 11:04:37.104159] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:31.621 [2024-10-29 11:04:37.104217] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.621 [2024-10-29 11:04:37.104240] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:31.621 [2024-10-29 11:04:37.104250] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.621 [2024-10-29 11:04:37.106472] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.621 [2024-10-29 11:04:37.106508] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:31.621 BaseBdev2 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.621 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.882 BaseBdev3_malloc 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.882 [2024-10-29 11:04:37.132994] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:31.882 [2024-10-29 11:04:37.133049] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.882 [2024-10-29 11:04:37.133075] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:31.882 [2024-10-29 11:04:37.133085] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.882 [2024-10-29 11:04:37.135187] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.882 [2024-10-29 11:04:37.135225] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:31.882 BaseBdev3 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.882 BaseBdev4_malloc 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.882 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.882 [2024-10-29 11:04:37.180177] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:31.883 [2024-10-29 11:04:37.180277] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.883 [2024-10-29 11:04:37.180327] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:31.883 [2024-10-29 11:04:37.180350] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.883 [2024-10-29 11:04:37.184664] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.883 [2024-10-29 11:04:37.184740] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:31.883 BaseBdev4 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.883 spare_malloc 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.883 spare_delay 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.883 [2024-10-29 11:04:37.222510] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:31.883 [2024-10-29 11:04:37.222626] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:31.883 [2024-10-29 11:04:37.222672] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:31.883 [2024-10-29 11:04:37.222723] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:31.883 [2024-10-29 11:04:37.224869] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:31.883 [2024-10-29 11:04:37.224947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:31.883 spare 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.883 [2024-10-29 11:04:37.234580] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:31.883 [2024-10-29 11:04:37.236385] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:31.883 [2024-10-29 11:04:37.236461] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:31.883 [2024-10-29 11:04:37.236504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:31.883 [2024-10-29 11:04:37.236705] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:15:31.883 [2024-10-29 11:04:37.236731] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:31.883 [2024-10-29 11:04:37.236974] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:15:31.883 [2024-10-29 11:04:37.237428] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:15:31.883 [2024-10-29 11:04:37.237444] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:15:31.883 [2024-10-29 11:04:37.237587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.883 "name": "raid_bdev1", 00:15:31.883 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:31.883 "strip_size_kb": 64, 00:15:31.883 "state": "online", 00:15:31.883 "raid_level": "raid5f", 00:15:31.883 "superblock": true, 00:15:31.883 "num_base_bdevs": 4, 00:15:31.883 "num_base_bdevs_discovered": 4, 00:15:31.883 "num_base_bdevs_operational": 4, 00:15:31.883 "base_bdevs_list": [ 00:15:31.883 { 00:15:31.883 "name": "BaseBdev1", 00:15:31.883 "uuid": "89826642-6d57-5abf-8a21-65a0b8e1faf8", 00:15:31.883 "is_configured": true, 00:15:31.883 "data_offset": 2048, 00:15:31.883 "data_size": 63488 00:15:31.883 }, 00:15:31.883 { 00:15:31.883 "name": "BaseBdev2", 00:15:31.883 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:31.883 "is_configured": true, 00:15:31.883 "data_offset": 2048, 00:15:31.883 "data_size": 63488 00:15:31.883 }, 00:15:31.883 { 00:15:31.883 "name": "BaseBdev3", 00:15:31.883 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:31.883 "is_configured": true, 00:15:31.883 "data_offset": 2048, 00:15:31.883 "data_size": 63488 00:15:31.883 }, 00:15:31.883 { 00:15:31.883 "name": "BaseBdev4", 00:15:31.883 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:31.883 "is_configured": true, 00:15:31.883 "data_offset": 2048, 00:15:31.883 "data_size": 63488 00:15:31.883 } 00:15:31.883 ] 00:15:31.883 }' 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.883 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.459 [2024-10-29 11:04:37.730686] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:32.459 11:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:32.719 [2024-10-29 11:04:37.998046] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:15:32.719 /dev/nbd0 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:32.719 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:32.719 1+0 records in 00:15:32.719 1+0 records out 00:15:32.720 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000354589 s, 11.6 MB/s 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:32.720 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:15:33.290 496+0 records in 00:15:33.290 496+0 records out 00:15:33.290 97517568 bytes (98 MB, 93 MiB) copied, 0.414032 s, 236 MB/s 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:33.290 [2024-10-29 11:04:38.677917] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.290 [2024-10-29 11:04:38.714005] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.290 "name": "raid_bdev1", 00:15:33.290 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:33.290 "strip_size_kb": 64, 00:15:33.290 "state": "online", 00:15:33.290 "raid_level": "raid5f", 00:15:33.290 "superblock": true, 00:15:33.290 "num_base_bdevs": 4, 00:15:33.290 "num_base_bdevs_discovered": 3, 00:15:33.290 "num_base_bdevs_operational": 3, 00:15:33.290 "base_bdevs_list": [ 00:15:33.290 { 00:15:33.290 "name": null, 00:15:33.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.290 "is_configured": false, 00:15:33.290 "data_offset": 0, 00:15:33.290 "data_size": 63488 00:15:33.290 }, 00:15:33.290 { 00:15:33.290 "name": "BaseBdev2", 00:15:33.290 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:33.290 "is_configured": true, 00:15:33.290 "data_offset": 2048, 00:15:33.290 "data_size": 63488 00:15:33.290 }, 00:15:33.290 { 00:15:33.290 "name": "BaseBdev3", 00:15:33.290 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:33.290 "is_configured": true, 00:15:33.290 "data_offset": 2048, 00:15:33.290 "data_size": 63488 00:15:33.290 }, 00:15:33.290 { 00:15:33.290 "name": "BaseBdev4", 00:15:33.290 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:33.290 "is_configured": true, 00:15:33.290 "data_offset": 2048, 00:15:33.290 "data_size": 63488 00:15:33.290 } 00:15:33.290 ] 00:15:33.290 }' 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.290 11:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.861 11:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:33.861 11:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.861 11:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:33.861 [2024-10-29 11:04:39.193234] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:33.861 [2024-10-29 11:04:39.197624] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002a8b0 00:15:33.861 11:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.861 11:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:33.861 [2024-10-29 11:04:39.199889] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.799 "name": "raid_bdev1", 00:15:34.799 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:34.799 "strip_size_kb": 64, 00:15:34.799 "state": "online", 00:15:34.799 "raid_level": "raid5f", 00:15:34.799 "superblock": true, 00:15:34.799 "num_base_bdevs": 4, 00:15:34.799 "num_base_bdevs_discovered": 4, 00:15:34.799 "num_base_bdevs_operational": 4, 00:15:34.799 "process": { 00:15:34.799 "type": "rebuild", 00:15:34.799 "target": "spare", 00:15:34.799 "progress": { 00:15:34.799 "blocks": 19200, 00:15:34.799 "percent": 10 00:15:34.799 } 00:15:34.799 }, 00:15:34.799 "base_bdevs_list": [ 00:15:34.799 { 00:15:34.799 "name": "spare", 00:15:34.799 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:34.799 "is_configured": true, 00:15:34.799 "data_offset": 2048, 00:15:34.799 "data_size": 63488 00:15:34.799 }, 00:15:34.799 { 00:15:34.799 "name": "BaseBdev2", 00:15:34.799 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:34.799 "is_configured": true, 00:15:34.799 "data_offset": 2048, 00:15:34.799 "data_size": 63488 00:15:34.799 }, 00:15:34.799 { 00:15:34.799 "name": "BaseBdev3", 00:15:34.799 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:34.799 "is_configured": true, 00:15:34.799 "data_offset": 2048, 00:15:34.799 "data_size": 63488 00:15:34.799 }, 00:15:34.799 { 00:15:34.799 "name": "BaseBdev4", 00:15:34.799 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:34.799 "is_configured": true, 00:15:34.799 "data_offset": 2048, 00:15:34.799 "data_size": 63488 00:15:34.799 } 00:15:34.799 ] 00:15:34.799 }' 00:15:34.799 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.060 [2024-10-29 11:04:40.360473] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:35.060 [2024-10-29 11:04:40.405379] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:35.060 [2024-10-29 11:04:40.405443] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:35.060 [2024-10-29 11:04:40.405462] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:35.060 [2024-10-29 11:04:40.405471] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:35.060 "name": "raid_bdev1", 00:15:35.060 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:35.060 "strip_size_kb": 64, 00:15:35.060 "state": "online", 00:15:35.060 "raid_level": "raid5f", 00:15:35.060 "superblock": true, 00:15:35.060 "num_base_bdevs": 4, 00:15:35.060 "num_base_bdevs_discovered": 3, 00:15:35.060 "num_base_bdevs_operational": 3, 00:15:35.060 "base_bdevs_list": [ 00:15:35.060 { 00:15:35.060 "name": null, 00:15:35.060 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:35.060 "is_configured": false, 00:15:35.060 "data_offset": 0, 00:15:35.060 "data_size": 63488 00:15:35.060 }, 00:15:35.060 { 00:15:35.060 "name": "BaseBdev2", 00:15:35.060 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:35.060 "is_configured": true, 00:15:35.060 "data_offset": 2048, 00:15:35.060 "data_size": 63488 00:15:35.060 }, 00:15:35.060 { 00:15:35.060 "name": "BaseBdev3", 00:15:35.060 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:35.060 "is_configured": true, 00:15:35.060 "data_offset": 2048, 00:15:35.060 "data_size": 63488 00:15:35.060 }, 00:15:35.060 { 00:15:35.060 "name": "BaseBdev4", 00:15:35.060 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:35.060 "is_configured": true, 00:15:35.060 "data_offset": 2048, 00:15:35.060 "data_size": 63488 00:15:35.060 } 00:15:35.060 ] 00:15:35.060 }' 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:35.060 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:35.631 "name": "raid_bdev1", 00:15:35.631 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:35.631 "strip_size_kb": 64, 00:15:35.631 "state": "online", 00:15:35.631 "raid_level": "raid5f", 00:15:35.631 "superblock": true, 00:15:35.631 "num_base_bdevs": 4, 00:15:35.631 "num_base_bdevs_discovered": 3, 00:15:35.631 "num_base_bdevs_operational": 3, 00:15:35.631 "base_bdevs_list": [ 00:15:35.631 { 00:15:35.631 "name": null, 00:15:35.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:35.631 "is_configured": false, 00:15:35.631 "data_offset": 0, 00:15:35.631 "data_size": 63488 00:15:35.631 }, 00:15:35.631 { 00:15:35.631 "name": "BaseBdev2", 00:15:35.631 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:35.631 "is_configured": true, 00:15:35.631 "data_offset": 2048, 00:15:35.631 "data_size": 63488 00:15:35.631 }, 00:15:35.631 { 00:15:35.631 "name": "BaseBdev3", 00:15:35.631 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:35.631 "is_configured": true, 00:15:35.631 "data_offset": 2048, 00:15:35.631 "data_size": 63488 00:15:35.631 }, 00:15:35.631 { 00:15:35.631 "name": "BaseBdev4", 00:15:35.631 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:35.631 "is_configured": true, 00:15:35.631 "data_offset": 2048, 00:15:35.631 "data_size": 63488 00:15:35.631 } 00:15:35.631 ] 00:15:35.631 }' 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:35.631 11:04:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.631 [2024-10-29 11:04:41.010012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:35.631 [2024-10-29 11:04:41.013832] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002a980 00:15:35.631 [2024-10-29 11:04:41.016094] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.631 11:04:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.572 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.832 "name": "raid_bdev1", 00:15:36.832 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:36.832 "strip_size_kb": 64, 00:15:36.832 "state": "online", 00:15:36.832 "raid_level": "raid5f", 00:15:36.832 "superblock": true, 00:15:36.832 "num_base_bdevs": 4, 00:15:36.832 "num_base_bdevs_discovered": 4, 00:15:36.832 "num_base_bdevs_operational": 4, 00:15:36.832 "process": { 00:15:36.832 "type": "rebuild", 00:15:36.832 "target": "spare", 00:15:36.832 "progress": { 00:15:36.832 "blocks": 19200, 00:15:36.832 "percent": 10 00:15:36.832 } 00:15:36.832 }, 00:15:36.832 "base_bdevs_list": [ 00:15:36.832 { 00:15:36.832 "name": "spare", 00:15:36.832 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev2", 00:15:36.832 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev3", 00:15:36.832 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev4", 00:15:36.832 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 } 00:15:36.832 ] 00:15:36.832 }' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:36.832 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=527 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:36.832 "name": "raid_bdev1", 00:15:36.832 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:36.832 "strip_size_kb": 64, 00:15:36.832 "state": "online", 00:15:36.832 "raid_level": "raid5f", 00:15:36.832 "superblock": true, 00:15:36.832 "num_base_bdevs": 4, 00:15:36.832 "num_base_bdevs_discovered": 4, 00:15:36.832 "num_base_bdevs_operational": 4, 00:15:36.832 "process": { 00:15:36.832 "type": "rebuild", 00:15:36.832 "target": "spare", 00:15:36.832 "progress": { 00:15:36.832 "blocks": 21120, 00:15:36.832 "percent": 11 00:15:36.832 } 00:15:36.832 }, 00:15:36.832 "base_bdevs_list": [ 00:15:36.832 { 00:15:36.832 "name": "spare", 00:15:36.832 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev2", 00:15:36.832 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev3", 00:15:36.832 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 }, 00:15:36.832 { 00:15:36.832 "name": "BaseBdev4", 00:15:36.832 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:36.832 "is_configured": true, 00:15:36.832 "data_offset": 2048, 00:15:36.832 "data_size": 63488 00:15:36.832 } 00:15:36.832 ] 00:15:36.832 }' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:36.832 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:36.833 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:36.833 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:36.833 11:04:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:38.215 "name": "raid_bdev1", 00:15:38.215 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:38.215 "strip_size_kb": 64, 00:15:38.215 "state": "online", 00:15:38.215 "raid_level": "raid5f", 00:15:38.215 "superblock": true, 00:15:38.215 "num_base_bdevs": 4, 00:15:38.215 "num_base_bdevs_discovered": 4, 00:15:38.215 "num_base_bdevs_operational": 4, 00:15:38.215 "process": { 00:15:38.215 "type": "rebuild", 00:15:38.215 "target": "spare", 00:15:38.215 "progress": { 00:15:38.215 "blocks": 44160, 00:15:38.215 "percent": 23 00:15:38.215 } 00:15:38.215 }, 00:15:38.215 "base_bdevs_list": [ 00:15:38.215 { 00:15:38.215 "name": "spare", 00:15:38.215 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:38.215 "is_configured": true, 00:15:38.215 "data_offset": 2048, 00:15:38.215 "data_size": 63488 00:15:38.215 }, 00:15:38.215 { 00:15:38.215 "name": "BaseBdev2", 00:15:38.215 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:38.215 "is_configured": true, 00:15:38.215 "data_offset": 2048, 00:15:38.215 "data_size": 63488 00:15:38.215 }, 00:15:38.215 { 00:15:38.215 "name": "BaseBdev3", 00:15:38.215 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:38.215 "is_configured": true, 00:15:38.215 "data_offset": 2048, 00:15:38.215 "data_size": 63488 00:15:38.215 }, 00:15:38.215 { 00:15:38.215 "name": "BaseBdev4", 00:15:38.215 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:38.215 "is_configured": true, 00:15:38.215 "data_offset": 2048, 00:15:38.215 "data_size": 63488 00:15:38.215 } 00:15:38.215 ] 00:15:38.215 }' 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:38.215 11:04:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:39.156 "name": "raid_bdev1", 00:15:39.156 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:39.156 "strip_size_kb": 64, 00:15:39.156 "state": "online", 00:15:39.156 "raid_level": "raid5f", 00:15:39.156 "superblock": true, 00:15:39.156 "num_base_bdevs": 4, 00:15:39.156 "num_base_bdevs_discovered": 4, 00:15:39.156 "num_base_bdevs_operational": 4, 00:15:39.156 "process": { 00:15:39.156 "type": "rebuild", 00:15:39.156 "target": "spare", 00:15:39.156 "progress": { 00:15:39.156 "blocks": 65280, 00:15:39.156 "percent": 34 00:15:39.156 } 00:15:39.156 }, 00:15:39.156 "base_bdevs_list": [ 00:15:39.156 { 00:15:39.156 "name": "spare", 00:15:39.156 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:39.156 "is_configured": true, 00:15:39.156 "data_offset": 2048, 00:15:39.156 "data_size": 63488 00:15:39.156 }, 00:15:39.156 { 00:15:39.156 "name": "BaseBdev2", 00:15:39.156 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:39.156 "is_configured": true, 00:15:39.156 "data_offset": 2048, 00:15:39.156 "data_size": 63488 00:15:39.156 }, 00:15:39.156 { 00:15:39.156 "name": "BaseBdev3", 00:15:39.156 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:39.156 "is_configured": true, 00:15:39.156 "data_offset": 2048, 00:15:39.156 "data_size": 63488 00:15:39.156 }, 00:15:39.156 { 00:15:39.156 "name": "BaseBdev4", 00:15:39.156 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:39.156 "is_configured": true, 00:15:39.156 "data_offset": 2048, 00:15:39.156 "data_size": 63488 00:15:39.156 } 00:15:39.156 ] 00:15:39.156 }' 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.156 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:39.157 11:04:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:40.541 "name": "raid_bdev1", 00:15:40.541 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:40.541 "strip_size_kb": 64, 00:15:40.541 "state": "online", 00:15:40.541 "raid_level": "raid5f", 00:15:40.541 "superblock": true, 00:15:40.541 "num_base_bdevs": 4, 00:15:40.541 "num_base_bdevs_discovered": 4, 00:15:40.541 "num_base_bdevs_operational": 4, 00:15:40.541 "process": { 00:15:40.541 "type": "rebuild", 00:15:40.541 "target": "spare", 00:15:40.541 "progress": { 00:15:40.541 "blocks": 88320, 00:15:40.541 "percent": 46 00:15:40.541 } 00:15:40.541 }, 00:15:40.541 "base_bdevs_list": [ 00:15:40.541 { 00:15:40.541 "name": "spare", 00:15:40.541 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:40.541 "is_configured": true, 00:15:40.541 "data_offset": 2048, 00:15:40.541 "data_size": 63488 00:15:40.541 }, 00:15:40.541 { 00:15:40.541 "name": "BaseBdev2", 00:15:40.541 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:40.541 "is_configured": true, 00:15:40.541 "data_offset": 2048, 00:15:40.541 "data_size": 63488 00:15:40.541 }, 00:15:40.541 { 00:15:40.541 "name": "BaseBdev3", 00:15:40.541 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:40.541 "is_configured": true, 00:15:40.541 "data_offset": 2048, 00:15:40.541 "data_size": 63488 00:15:40.541 }, 00:15:40.541 { 00:15:40.541 "name": "BaseBdev4", 00:15:40.541 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:40.541 "is_configured": true, 00:15:40.541 "data_offset": 2048, 00:15:40.541 "data_size": 63488 00:15:40.541 } 00:15:40.541 ] 00:15:40.541 }' 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:40.541 11:04:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:41.482 "name": "raid_bdev1", 00:15:41.482 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:41.482 "strip_size_kb": 64, 00:15:41.482 "state": "online", 00:15:41.482 "raid_level": "raid5f", 00:15:41.482 "superblock": true, 00:15:41.482 "num_base_bdevs": 4, 00:15:41.482 "num_base_bdevs_discovered": 4, 00:15:41.482 "num_base_bdevs_operational": 4, 00:15:41.482 "process": { 00:15:41.482 "type": "rebuild", 00:15:41.482 "target": "spare", 00:15:41.482 "progress": { 00:15:41.482 "blocks": 109440, 00:15:41.482 "percent": 57 00:15:41.482 } 00:15:41.482 }, 00:15:41.482 "base_bdevs_list": [ 00:15:41.482 { 00:15:41.482 "name": "spare", 00:15:41.482 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:41.482 "is_configured": true, 00:15:41.482 "data_offset": 2048, 00:15:41.482 "data_size": 63488 00:15:41.482 }, 00:15:41.482 { 00:15:41.482 "name": "BaseBdev2", 00:15:41.482 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:41.482 "is_configured": true, 00:15:41.482 "data_offset": 2048, 00:15:41.482 "data_size": 63488 00:15:41.482 }, 00:15:41.482 { 00:15:41.482 "name": "BaseBdev3", 00:15:41.482 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:41.482 "is_configured": true, 00:15:41.482 "data_offset": 2048, 00:15:41.482 "data_size": 63488 00:15:41.482 }, 00:15:41.482 { 00:15:41.482 "name": "BaseBdev4", 00:15:41.482 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:41.482 "is_configured": true, 00:15:41.482 "data_offset": 2048, 00:15:41.482 "data_size": 63488 00:15:41.482 } 00:15:41.482 ] 00:15:41.482 }' 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:41.482 11:04:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:42.865 "name": "raid_bdev1", 00:15:42.865 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:42.865 "strip_size_kb": 64, 00:15:42.865 "state": "online", 00:15:42.865 "raid_level": "raid5f", 00:15:42.865 "superblock": true, 00:15:42.865 "num_base_bdevs": 4, 00:15:42.865 "num_base_bdevs_discovered": 4, 00:15:42.865 "num_base_bdevs_operational": 4, 00:15:42.865 "process": { 00:15:42.865 "type": "rebuild", 00:15:42.865 "target": "spare", 00:15:42.865 "progress": { 00:15:42.865 "blocks": 132480, 00:15:42.865 "percent": 69 00:15:42.865 } 00:15:42.865 }, 00:15:42.865 "base_bdevs_list": [ 00:15:42.865 { 00:15:42.865 "name": "spare", 00:15:42.865 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:42.865 "is_configured": true, 00:15:42.865 "data_offset": 2048, 00:15:42.865 "data_size": 63488 00:15:42.865 }, 00:15:42.865 { 00:15:42.865 "name": "BaseBdev2", 00:15:42.865 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:42.865 "is_configured": true, 00:15:42.865 "data_offset": 2048, 00:15:42.865 "data_size": 63488 00:15:42.865 }, 00:15:42.865 { 00:15:42.865 "name": "BaseBdev3", 00:15:42.865 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:42.865 "is_configured": true, 00:15:42.865 "data_offset": 2048, 00:15:42.865 "data_size": 63488 00:15:42.865 }, 00:15:42.865 { 00:15:42.865 "name": "BaseBdev4", 00:15:42.865 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:42.865 "is_configured": true, 00:15:42.865 "data_offset": 2048, 00:15:42.865 "data_size": 63488 00:15:42.865 } 00:15:42.865 ] 00:15:42.865 }' 00:15:42.865 11:04:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:42.865 11:04:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:42.865 11:04:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:42.865 11:04:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:42.865 11:04:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.805 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.805 "name": "raid_bdev1", 00:15:43.805 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:43.805 "strip_size_kb": 64, 00:15:43.805 "state": "online", 00:15:43.805 "raid_level": "raid5f", 00:15:43.805 "superblock": true, 00:15:43.805 "num_base_bdevs": 4, 00:15:43.805 "num_base_bdevs_discovered": 4, 00:15:43.805 "num_base_bdevs_operational": 4, 00:15:43.805 "process": { 00:15:43.805 "type": "rebuild", 00:15:43.805 "target": "spare", 00:15:43.805 "progress": { 00:15:43.805 "blocks": 153600, 00:15:43.805 "percent": 80 00:15:43.805 } 00:15:43.805 }, 00:15:43.805 "base_bdevs_list": [ 00:15:43.805 { 00:15:43.805 "name": "spare", 00:15:43.805 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:43.806 "is_configured": true, 00:15:43.806 "data_offset": 2048, 00:15:43.806 "data_size": 63488 00:15:43.806 }, 00:15:43.806 { 00:15:43.806 "name": "BaseBdev2", 00:15:43.806 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:43.806 "is_configured": true, 00:15:43.806 "data_offset": 2048, 00:15:43.806 "data_size": 63488 00:15:43.806 }, 00:15:43.806 { 00:15:43.806 "name": "BaseBdev3", 00:15:43.806 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:43.806 "is_configured": true, 00:15:43.806 "data_offset": 2048, 00:15:43.806 "data_size": 63488 00:15:43.806 }, 00:15:43.806 { 00:15:43.806 "name": "BaseBdev4", 00:15:43.806 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:43.806 "is_configured": true, 00:15:43.806 "data_offset": 2048, 00:15:43.806 "data_size": 63488 00:15:43.806 } 00:15:43.806 ] 00:15:43.806 }' 00:15:43.806 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:43.806 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:43.806 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:43.806 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:43.806 11:04:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:44.756 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:45.016 "name": "raid_bdev1", 00:15:45.016 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:45.016 "strip_size_kb": 64, 00:15:45.016 "state": "online", 00:15:45.016 "raid_level": "raid5f", 00:15:45.016 "superblock": true, 00:15:45.016 "num_base_bdevs": 4, 00:15:45.016 "num_base_bdevs_discovered": 4, 00:15:45.016 "num_base_bdevs_operational": 4, 00:15:45.016 "process": { 00:15:45.016 "type": "rebuild", 00:15:45.016 "target": "spare", 00:15:45.016 "progress": { 00:15:45.016 "blocks": 174720, 00:15:45.016 "percent": 91 00:15:45.016 } 00:15:45.016 }, 00:15:45.016 "base_bdevs_list": [ 00:15:45.016 { 00:15:45.016 "name": "spare", 00:15:45.016 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:45.016 "is_configured": true, 00:15:45.016 "data_offset": 2048, 00:15:45.016 "data_size": 63488 00:15:45.016 }, 00:15:45.016 { 00:15:45.016 "name": "BaseBdev2", 00:15:45.016 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:45.016 "is_configured": true, 00:15:45.016 "data_offset": 2048, 00:15:45.016 "data_size": 63488 00:15:45.016 }, 00:15:45.016 { 00:15:45.016 "name": "BaseBdev3", 00:15:45.016 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:45.016 "is_configured": true, 00:15:45.016 "data_offset": 2048, 00:15:45.016 "data_size": 63488 00:15:45.016 }, 00:15:45.016 { 00:15:45.016 "name": "BaseBdev4", 00:15:45.016 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:45.016 "is_configured": true, 00:15:45.016 "data_offset": 2048, 00:15:45.016 "data_size": 63488 00:15:45.016 } 00:15:45.016 ] 00:15:45.016 }' 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:45.016 11:04:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:45.590 [2024-10-29 11:04:51.056672] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:45.590 [2024-10-29 11:04:51.056822] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:45.590 [2024-10-29 11:04:51.056977] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.213 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.213 "name": "raid_bdev1", 00:15:46.213 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:46.213 "strip_size_kb": 64, 00:15:46.213 "state": "online", 00:15:46.213 "raid_level": "raid5f", 00:15:46.213 "superblock": true, 00:15:46.213 "num_base_bdevs": 4, 00:15:46.213 "num_base_bdevs_discovered": 4, 00:15:46.213 "num_base_bdevs_operational": 4, 00:15:46.213 "base_bdevs_list": [ 00:15:46.213 { 00:15:46.213 "name": "spare", 00:15:46.213 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:46.213 "is_configured": true, 00:15:46.213 "data_offset": 2048, 00:15:46.213 "data_size": 63488 00:15:46.213 }, 00:15:46.213 { 00:15:46.214 "name": "BaseBdev2", 00:15:46.214 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 }, 00:15:46.214 { 00:15:46.214 "name": "BaseBdev3", 00:15:46.214 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 }, 00:15:46.214 { 00:15:46.214 "name": "BaseBdev4", 00:15:46.214 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 } 00:15:46.214 ] 00:15:46.214 }' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.214 "name": "raid_bdev1", 00:15:46.214 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:46.214 "strip_size_kb": 64, 00:15:46.214 "state": "online", 00:15:46.214 "raid_level": "raid5f", 00:15:46.214 "superblock": true, 00:15:46.214 "num_base_bdevs": 4, 00:15:46.214 "num_base_bdevs_discovered": 4, 00:15:46.214 "num_base_bdevs_operational": 4, 00:15:46.214 "base_bdevs_list": [ 00:15:46.214 { 00:15:46.214 "name": "spare", 00:15:46.214 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 }, 00:15:46.214 { 00:15:46.214 "name": "BaseBdev2", 00:15:46.214 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 }, 00:15:46.214 { 00:15:46.214 "name": "BaseBdev3", 00:15:46.214 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 }, 00:15:46.214 { 00:15:46.214 "name": "BaseBdev4", 00:15:46.214 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:46.214 "is_configured": true, 00:15:46.214 "data_offset": 2048, 00:15:46.214 "data_size": 63488 00:15:46.214 } 00:15:46.214 ] 00:15:46.214 }' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.214 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.490 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.490 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.490 "name": "raid_bdev1", 00:15:46.490 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:46.490 "strip_size_kb": 64, 00:15:46.490 "state": "online", 00:15:46.490 "raid_level": "raid5f", 00:15:46.490 "superblock": true, 00:15:46.490 "num_base_bdevs": 4, 00:15:46.490 "num_base_bdevs_discovered": 4, 00:15:46.490 "num_base_bdevs_operational": 4, 00:15:46.490 "base_bdevs_list": [ 00:15:46.490 { 00:15:46.490 "name": "spare", 00:15:46.490 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:46.490 "is_configured": true, 00:15:46.490 "data_offset": 2048, 00:15:46.490 "data_size": 63488 00:15:46.490 }, 00:15:46.490 { 00:15:46.490 "name": "BaseBdev2", 00:15:46.490 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:46.490 "is_configured": true, 00:15:46.490 "data_offset": 2048, 00:15:46.490 "data_size": 63488 00:15:46.490 }, 00:15:46.490 { 00:15:46.490 "name": "BaseBdev3", 00:15:46.490 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:46.490 "is_configured": true, 00:15:46.490 "data_offset": 2048, 00:15:46.490 "data_size": 63488 00:15:46.490 }, 00:15:46.490 { 00:15:46.490 "name": "BaseBdev4", 00:15:46.490 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:46.490 "is_configured": true, 00:15:46.490 "data_offset": 2048, 00:15:46.490 "data_size": 63488 00:15:46.490 } 00:15:46.490 ] 00:15:46.490 }' 00:15:46.490 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.490 11:04:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.750 [2024-10-29 11:04:52.160795] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:46.750 [2024-10-29 11:04:52.160884] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:46.750 [2024-10-29 11:04:52.160983] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:46.750 [2024-10-29 11:04:52.161146] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:46.750 [2024-10-29 11:04:52.161209] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:46.750 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:46.751 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:47.011 /dev/nbd0 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:47.011 1+0 records in 00:15:47.011 1+0 records out 00:15:47.011 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000546277 s, 7.5 MB/s 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:47.011 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:47.272 /dev/nbd1 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # local i 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # break 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:47.272 1+0 records in 00:15:47.272 1+0 records out 00:15:47.272 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000405564 s, 10.1 MB/s 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # size=4096 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # return 0 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:47.272 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:47.533 11:04:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:47.533 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.793 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:47.793 [2024-10-29 11:04:53.266643] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:47.793 [2024-10-29 11:04:53.266710] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:47.793 [2024-10-29 11:04:53.266734] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:15:47.793 [2024-10-29 11:04:53.266747] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:47.793 [2024-10-29 11:04:53.268970] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:47.793 [2024-10-29 11:04:53.269036] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:47.793 [2024-10-29 11:04:53.269123] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:47.793 [2024-10-29 11:04:53.269167] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:47.793 [2024-10-29 11:04:53.269294] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:47.793 [2024-10-29 11:04:53.269407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:47.793 [2024-10-29 11:04:53.269481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:47.793 spare 00:15:47.794 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.794 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:47.794 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.794 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.054 [2024-10-29 11:04:53.369407] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:15:48.054 [2024-10-29 11:04:53.369436] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:48.054 [2024-10-29 11:04:53.369671] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049030 00:15:48.054 [2024-10-29 11:04:53.370099] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:15:48.054 [2024-10-29 11:04:53.370114] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:15:48.054 [2024-10-29 11:04:53.370239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.054 "name": "raid_bdev1", 00:15:48.054 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:48.054 "strip_size_kb": 64, 00:15:48.054 "state": "online", 00:15:48.054 "raid_level": "raid5f", 00:15:48.054 "superblock": true, 00:15:48.054 "num_base_bdevs": 4, 00:15:48.054 "num_base_bdevs_discovered": 4, 00:15:48.054 "num_base_bdevs_operational": 4, 00:15:48.054 "base_bdevs_list": [ 00:15:48.054 { 00:15:48.054 "name": "spare", 00:15:48.054 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:48.054 "is_configured": true, 00:15:48.054 "data_offset": 2048, 00:15:48.054 "data_size": 63488 00:15:48.054 }, 00:15:48.054 { 00:15:48.054 "name": "BaseBdev2", 00:15:48.054 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:48.054 "is_configured": true, 00:15:48.054 "data_offset": 2048, 00:15:48.054 "data_size": 63488 00:15:48.054 }, 00:15:48.054 { 00:15:48.054 "name": "BaseBdev3", 00:15:48.054 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:48.054 "is_configured": true, 00:15:48.054 "data_offset": 2048, 00:15:48.054 "data_size": 63488 00:15:48.054 }, 00:15:48.054 { 00:15:48.054 "name": "BaseBdev4", 00:15:48.054 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:48.054 "is_configured": true, 00:15:48.054 "data_offset": 2048, 00:15:48.054 "data_size": 63488 00:15:48.054 } 00:15:48.054 ] 00:15:48.054 }' 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.054 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.626 "name": "raid_bdev1", 00:15:48.626 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:48.626 "strip_size_kb": 64, 00:15:48.626 "state": "online", 00:15:48.626 "raid_level": "raid5f", 00:15:48.626 "superblock": true, 00:15:48.626 "num_base_bdevs": 4, 00:15:48.626 "num_base_bdevs_discovered": 4, 00:15:48.626 "num_base_bdevs_operational": 4, 00:15:48.626 "base_bdevs_list": [ 00:15:48.626 { 00:15:48.626 "name": "spare", 00:15:48.626 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev2", 00:15:48.626 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev3", 00:15:48.626 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev4", 00:15:48.626 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 } 00:15:48.626 ] 00:15:48.626 }' 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:48.626 11:04:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.626 [2024-10-29 11:04:54.034736] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.626 "name": "raid_bdev1", 00:15:48.626 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:48.626 "strip_size_kb": 64, 00:15:48.626 "state": "online", 00:15:48.626 "raid_level": "raid5f", 00:15:48.626 "superblock": true, 00:15:48.626 "num_base_bdevs": 4, 00:15:48.626 "num_base_bdevs_discovered": 3, 00:15:48.626 "num_base_bdevs_operational": 3, 00:15:48.626 "base_bdevs_list": [ 00:15:48.626 { 00:15:48.626 "name": null, 00:15:48.626 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.626 "is_configured": false, 00:15:48.626 "data_offset": 0, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev2", 00:15:48.626 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev3", 00:15:48.626 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 }, 00:15:48.626 { 00:15:48.626 "name": "BaseBdev4", 00:15:48.626 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:48.626 "is_configured": true, 00:15:48.626 "data_offset": 2048, 00:15:48.626 "data_size": 63488 00:15:48.626 } 00:15:48.626 ] 00:15:48.626 }' 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.626 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.198 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:49.198 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.198 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.198 [2024-10-29 11:04:54.509950] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:49.198 [2024-10-29 11:04:54.510233] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:49.198 [2024-10-29 11:04:54.510305] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:49.198 [2024-10-29 11:04:54.510405] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:49.198 [2024-10-29 11:04:54.514604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049100 00:15:49.198 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.198 11:04:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:49.198 [2024-10-29 11:04:54.516937] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:50.139 "name": "raid_bdev1", 00:15:50.139 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:50.139 "strip_size_kb": 64, 00:15:50.139 "state": "online", 00:15:50.139 "raid_level": "raid5f", 00:15:50.139 "superblock": true, 00:15:50.139 "num_base_bdevs": 4, 00:15:50.139 "num_base_bdevs_discovered": 4, 00:15:50.139 "num_base_bdevs_operational": 4, 00:15:50.139 "process": { 00:15:50.139 "type": "rebuild", 00:15:50.139 "target": "spare", 00:15:50.139 "progress": { 00:15:50.139 "blocks": 19200, 00:15:50.139 "percent": 10 00:15:50.139 } 00:15:50.139 }, 00:15:50.139 "base_bdevs_list": [ 00:15:50.139 { 00:15:50.139 "name": "spare", 00:15:50.139 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:50.139 "is_configured": true, 00:15:50.139 "data_offset": 2048, 00:15:50.139 "data_size": 63488 00:15:50.139 }, 00:15:50.139 { 00:15:50.139 "name": "BaseBdev2", 00:15:50.139 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:50.139 "is_configured": true, 00:15:50.139 "data_offset": 2048, 00:15:50.139 "data_size": 63488 00:15:50.139 }, 00:15:50.139 { 00:15:50.139 "name": "BaseBdev3", 00:15:50.139 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:50.139 "is_configured": true, 00:15:50.139 "data_offset": 2048, 00:15:50.139 "data_size": 63488 00:15:50.139 }, 00:15:50.139 { 00:15:50.139 "name": "BaseBdev4", 00:15:50.139 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:50.139 "is_configured": true, 00:15:50.139 "data_offset": 2048, 00:15:50.139 "data_size": 63488 00:15:50.139 } 00:15:50.139 ] 00:15:50.139 }' 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:50.139 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.400 [2024-10-29 11:04:55.680666] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:50.400 [2024-10-29 11:04:55.722215] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:50.400 [2024-10-29 11:04:55.722274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:50.400 [2024-10-29 11:04:55.722295] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:50.400 [2024-10-29 11:04:55.722304] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.400 "name": "raid_bdev1", 00:15:50.400 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:50.400 "strip_size_kb": 64, 00:15:50.400 "state": "online", 00:15:50.400 "raid_level": "raid5f", 00:15:50.400 "superblock": true, 00:15:50.400 "num_base_bdevs": 4, 00:15:50.400 "num_base_bdevs_discovered": 3, 00:15:50.400 "num_base_bdevs_operational": 3, 00:15:50.400 "base_bdevs_list": [ 00:15:50.400 { 00:15:50.400 "name": null, 00:15:50.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:50.400 "is_configured": false, 00:15:50.400 "data_offset": 0, 00:15:50.400 "data_size": 63488 00:15:50.400 }, 00:15:50.400 { 00:15:50.400 "name": "BaseBdev2", 00:15:50.400 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:50.400 "is_configured": true, 00:15:50.400 "data_offset": 2048, 00:15:50.400 "data_size": 63488 00:15:50.400 }, 00:15:50.400 { 00:15:50.400 "name": "BaseBdev3", 00:15:50.400 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:50.400 "is_configured": true, 00:15:50.400 "data_offset": 2048, 00:15:50.400 "data_size": 63488 00:15:50.400 }, 00:15:50.400 { 00:15:50.400 "name": "BaseBdev4", 00:15:50.400 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:50.400 "is_configured": true, 00:15:50.400 "data_offset": 2048, 00:15:50.400 "data_size": 63488 00:15:50.400 } 00:15:50.400 ] 00:15:50.400 }' 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.400 11:04:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.971 11:04:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:50.971 11:04:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.971 11:04:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.971 [2024-10-29 11:04:56.206689] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:50.971 [2024-10-29 11:04:56.206814] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:50.971 [2024-10-29 11:04:56.206861] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:15:50.971 [2024-10-29 11:04:56.206894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:50.971 [2024-10-29 11:04:56.207422] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:50.971 [2024-10-29 11:04:56.207496] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:50.971 [2024-10-29 11:04:56.207635] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:50.971 [2024-10-29 11:04:56.207678] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:50.971 [2024-10-29 11:04:56.207760] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:50.971 [2024-10-29 11:04:56.207845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:50.971 [2024-10-29 11:04:56.211228] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000491d0 00:15:50.971 spare 00:15:50.971 11:04:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.971 11:04:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:50.971 [2024-10-29 11:04:56.213530] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:51.912 "name": "raid_bdev1", 00:15:51.912 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:51.912 "strip_size_kb": 64, 00:15:51.912 "state": "online", 00:15:51.912 "raid_level": "raid5f", 00:15:51.912 "superblock": true, 00:15:51.912 "num_base_bdevs": 4, 00:15:51.912 "num_base_bdevs_discovered": 4, 00:15:51.912 "num_base_bdevs_operational": 4, 00:15:51.912 "process": { 00:15:51.912 "type": "rebuild", 00:15:51.912 "target": "spare", 00:15:51.912 "progress": { 00:15:51.912 "blocks": 19200, 00:15:51.912 "percent": 10 00:15:51.912 } 00:15:51.912 }, 00:15:51.912 "base_bdevs_list": [ 00:15:51.912 { 00:15:51.912 "name": "spare", 00:15:51.912 "uuid": "d4044b83-28c9-5343-98e5-1cfef7bc170a", 00:15:51.912 "is_configured": true, 00:15:51.912 "data_offset": 2048, 00:15:51.912 "data_size": 63488 00:15:51.912 }, 00:15:51.912 { 00:15:51.912 "name": "BaseBdev2", 00:15:51.912 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:51.912 "is_configured": true, 00:15:51.912 "data_offset": 2048, 00:15:51.912 "data_size": 63488 00:15:51.912 }, 00:15:51.912 { 00:15:51.912 "name": "BaseBdev3", 00:15:51.912 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:51.912 "is_configured": true, 00:15:51.912 "data_offset": 2048, 00:15:51.912 "data_size": 63488 00:15:51.912 }, 00:15:51.912 { 00:15:51.912 "name": "BaseBdev4", 00:15:51.912 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:51.912 "is_configured": true, 00:15:51.912 "data_offset": 2048, 00:15:51.912 "data_size": 63488 00:15:51.912 } 00:15:51.912 ] 00:15:51.912 }' 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.912 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.912 [2024-10-29 11:04:57.377813] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:52.173 [2024-10-29 11:04:57.418834] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:52.173 [2024-10-29 11:04:57.418892] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:52.173 [2024-10-29 11:04:57.418910] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:52.173 [2024-10-29 11:04:57.418922] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.173 "name": "raid_bdev1", 00:15:52.173 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:52.173 "strip_size_kb": 64, 00:15:52.173 "state": "online", 00:15:52.173 "raid_level": "raid5f", 00:15:52.173 "superblock": true, 00:15:52.173 "num_base_bdevs": 4, 00:15:52.173 "num_base_bdevs_discovered": 3, 00:15:52.173 "num_base_bdevs_operational": 3, 00:15:52.173 "base_bdevs_list": [ 00:15:52.173 { 00:15:52.173 "name": null, 00:15:52.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.173 "is_configured": false, 00:15:52.173 "data_offset": 0, 00:15:52.173 "data_size": 63488 00:15:52.173 }, 00:15:52.173 { 00:15:52.173 "name": "BaseBdev2", 00:15:52.173 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:52.173 "is_configured": true, 00:15:52.173 "data_offset": 2048, 00:15:52.173 "data_size": 63488 00:15:52.173 }, 00:15:52.173 { 00:15:52.173 "name": "BaseBdev3", 00:15:52.173 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:52.173 "is_configured": true, 00:15:52.173 "data_offset": 2048, 00:15:52.173 "data_size": 63488 00:15:52.173 }, 00:15:52.173 { 00:15:52.173 "name": "BaseBdev4", 00:15:52.173 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:52.173 "is_configured": true, 00:15:52.173 "data_offset": 2048, 00:15:52.173 "data_size": 63488 00:15:52.173 } 00:15:52.173 ] 00:15:52.173 }' 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.173 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.433 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:52.433 "name": "raid_bdev1", 00:15:52.433 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:52.433 "strip_size_kb": 64, 00:15:52.433 "state": "online", 00:15:52.433 "raid_level": "raid5f", 00:15:52.433 "superblock": true, 00:15:52.433 "num_base_bdevs": 4, 00:15:52.433 "num_base_bdevs_discovered": 3, 00:15:52.433 "num_base_bdevs_operational": 3, 00:15:52.433 "base_bdevs_list": [ 00:15:52.433 { 00:15:52.433 "name": null, 00:15:52.434 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.434 "is_configured": false, 00:15:52.434 "data_offset": 0, 00:15:52.434 "data_size": 63488 00:15:52.434 }, 00:15:52.434 { 00:15:52.434 "name": "BaseBdev2", 00:15:52.434 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:52.434 "is_configured": true, 00:15:52.434 "data_offset": 2048, 00:15:52.434 "data_size": 63488 00:15:52.434 }, 00:15:52.434 { 00:15:52.434 "name": "BaseBdev3", 00:15:52.434 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:52.434 "is_configured": true, 00:15:52.434 "data_offset": 2048, 00:15:52.434 "data_size": 63488 00:15:52.434 }, 00:15:52.434 { 00:15:52.434 "name": "BaseBdev4", 00:15:52.434 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:52.434 "is_configured": true, 00:15:52.434 "data_offset": 2048, 00:15:52.434 "data_size": 63488 00:15:52.434 } 00:15:52.434 ] 00:15:52.434 }' 00:15:52.434 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.694 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:52.694 11:04:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.694 [2024-10-29 11:04:58.031651] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:52.694 [2024-10-29 11:04:58.031767] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:52.694 [2024-10-29 11:04:58.031807] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:15:52.694 [2024-10-29 11:04:58.031841] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:52.694 [2024-10-29 11:04:58.032353] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:52.694 [2024-10-29 11:04:58.032446] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:52.694 [2024-10-29 11:04:58.032569] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:52.694 [2024-10-29 11:04:58.032622] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:52.694 [2024-10-29 11:04:58.032690] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:52.694 [2024-10-29 11:04:58.032736] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:52.694 BaseBdev1 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.694 11:04:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.634 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.634 "name": "raid_bdev1", 00:15:53.634 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:53.634 "strip_size_kb": 64, 00:15:53.634 "state": "online", 00:15:53.634 "raid_level": "raid5f", 00:15:53.634 "superblock": true, 00:15:53.634 "num_base_bdevs": 4, 00:15:53.634 "num_base_bdevs_discovered": 3, 00:15:53.635 "num_base_bdevs_operational": 3, 00:15:53.635 "base_bdevs_list": [ 00:15:53.635 { 00:15:53.635 "name": null, 00:15:53.635 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.635 "is_configured": false, 00:15:53.635 "data_offset": 0, 00:15:53.635 "data_size": 63488 00:15:53.635 }, 00:15:53.635 { 00:15:53.635 "name": "BaseBdev2", 00:15:53.635 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:53.635 "is_configured": true, 00:15:53.635 "data_offset": 2048, 00:15:53.635 "data_size": 63488 00:15:53.635 }, 00:15:53.635 { 00:15:53.635 "name": "BaseBdev3", 00:15:53.635 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:53.635 "is_configured": true, 00:15:53.635 "data_offset": 2048, 00:15:53.635 "data_size": 63488 00:15:53.635 }, 00:15:53.635 { 00:15:53.635 "name": "BaseBdev4", 00:15:53.635 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:53.635 "is_configured": true, 00:15:53.635 "data_offset": 2048, 00:15:53.635 "data_size": 63488 00:15:53.635 } 00:15:53.635 ] 00:15:53.635 }' 00:15:53.635 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.635 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.205 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:54.205 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:54.205 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:54.205 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:54.205 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:54.206 "name": "raid_bdev1", 00:15:54.206 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:54.206 "strip_size_kb": 64, 00:15:54.206 "state": "online", 00:15:54.206 "raid_level": "raid5f", 00:15:54.206 "superblock": true, 00:15:54.206 "num_base_bdevs": 4, 00:15:54.206 "num_base_bdevs_discovered": 3, 00:15:54.206 "num_base_bdevs_operational": 3, 00:15:54.206 "base_bdevs_list": [ 00:15:54.206 { 00:15:54.206 "name": null, 00:15:54.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.206 "is_configured": false, 00:15:54.206 "data_offset": 0, 00:15:54.206 "data_size": 63488 00:15:54.206 }, 00:15:54.206 { 00:15:54.206 "name": "BaseBdev2", 00:15:54.206 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:54.206 "is_configured": true, 00:15:54.206 "data_offset": 2048, 00:15:54.206 "data_size": 63488 00:15:54.206 }, 00:15:54.206 { 00:15:54.206 "name": "BaseBdev3", 00:15:54.206 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:54.206 "is_configured": true, 00:15:54.206 "data_offset": 2048, 00:15:54.206 "data_size": 63488 00:15:54.206 }, 00:15:54.206 { 00:15:54.206 "name": "BaseBdev4", 00:15:54.206 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:54.206 "is_configured": true, 00:15:54.206 "data_offset": 2048, 00:15:54.206 "data_size": 63488 00:15:54.206 } 00:15:54.206 ] 00:15:54.206 }' 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.206 [2024-10-29 11:04:59.652922] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:54.206 [2024-10-29 11:04:59.653059] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:54.206 [2024-10-29 11:04:59.653076] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:54.206 request: 00:15:54.206 { 00:15:54.206 "base_bdev": "BaseBdev1", 00:15:54.206 "raid_bdev": "raid_bdev1", 00:15:54.206 "method": "bdev_raid_add_base_bdev", 00:15:54.206 "req_id": 1 00:15:54.206 } 00:15:54.206 Got JSON-RPC error response 00:15:54.206 response: 00:15:54.206 { 00:15:54.206 "code": -22, 00:15:54.206 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:54.206 } 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:54.206 11:04:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.588 "name": "raid_bdev1", 00:15:55.588 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:55.588 "strip_size_kb": 64, 00:15:55.588 "state": "online", 00:15:55.588 "raid_level": "raid5f", 00:15:55.588 "superblock": true, 00:15:55.588 "num_base_bdevs": 4, 00:15:55.588 "num_base_bdevs_discovered": 3, 00:15:55.588 "num_base_bdevs_operational": 3, 00:15:55.588 "base_bdevs_list": [ 00:15:55.588 { 00:15:55.588 "name": null, 00:15:55.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.588 "is_configured": false, 00:15:55.588 "data_offset": 0, 00:15:55.588 "data_size": 63488 00:15:55.588 }, 00:15:55.588 { 00:15:55.588 "name": "BaseBdev2", 00:15:55.588 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:55.588 "is_configured": true, 00:15:55.588 "data_offset": 2048, 00:15:55.588 "data_size": 63488 00:15:55.588 }, 00:15:55.588 { 00:15:55.588 "name": "BaseBdev3", 00:15:55.588 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:55.588 "is_configured": true, 00:15:55.588 "data_offset": 2048, 00:15:55.588 "data_size": 63488 00:15:55.588 }, 00:15:55.588 { 00:15:55.588 "name": "BaseBdev4", 00:15:55.588 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:55.588 "is_configured": true, 00:15:55.588 "data_offset": 2048, 00:15:55.588 "data_size": 63488 00:15:55.588 } 00:15:55.588 ] 00:15:55.588 }' 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.588 11:05:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.848 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:55.849 "name": "raid_bdev1", 00:15:55.849 "uuid": "2cd8ee6c-46fa-4f08-9e41-019d9e82bfd4", 00:15:55.849 "strip_size_kb": 64, 00:15:55.849 "state": "online", 00:15:55.849 "raid_level": "raid5f", 00:15:55.849 "superblock": true, 00:15:55.849 "num_base_bdevs": 4, 00:15:55.849 "num_base_bdevs_discovered": 3, 00:15:55.849 "num_base_bdevs_operational": 3, 00:15:55.849 "base_bdevs_list": [ 00:15:55.849 { 00:15:55.849 "name": null, 00:15:55.849 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.849 "is_configured": false, 00:15:55.849 "data_offset": 0, 00:15:55.849 "data_size": 63488 00:15:55.849 }, 00:15:55.849 { 00:15:55.849 "name": "BaseBdev2", 00:15:55.849 "uuid": "4be6493f-7561-5132-83f5-a7f446a70267", 00:15:55.849 "is_configured": true, 00:15:55.849 "data_offset": 2048, 00:15:55.849 "data_size": 63488 00:15:55.849 }, 00:15:55.849 { 00:15:55.849 "name": "BaseBdev3", 00:15:55.849 "uuid": "f0c981a2-bc25-5556-98d7-0a10c6ad73ba", 00:15:55.849 "is_configured": true, 00:15:55.849 "data_offset": 2048, 00:15:55.849 "data_size": 63488 00:15:55.849 }, 00:15:55.849 { 00:15:55.849 "name": "BaseBdev4", 00:15:55.849 "uuid": "479b099a-7be2-5c3a-b973-1a6ca93b931c", 00:15:55.849 "is_configured": true, 00:15:55.849 "data_offset": 2048, 00:15:55.849 "data_size": 63488 00:15:55.849 } 00:15:55.849 ] 00:15:55.849 }' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95621 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@952 -- # '[' -z 95621 ']' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # kill -0 95621 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@957 -- # uname 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 95621 00:15:55.849 killing process with pid 95621 00:15:55.849 Received shutdown signal, test time was about 60.000000 seconds 00:15:55.849 00:15:55.849 Latency(us) 00:15:55.849 [2024-10-29T11:05:01.347Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:55.849 [2024-10-29T11:05:01.347Z] =================================================================================================================== 00:15:55.849 [2024-10-29T11:05:01.347Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@970 -- # echo 'killing process with pid 95621' 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@971 -- # kill 95621 00:15:55.849 [2024-10-29 11:05:01.280562] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:55.849 [2024-10-29 11:05:01.280668] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:55.849 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@976 -- # wait 95621 00:15:55.849 [2024-10-29 11:05:01.280745] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:55.849 [2024-10-29 11:05:01.280757] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:15:55.849 [2024-10-29 11:05:01.331379] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:56.110 ************************************ 00:15:56.110 END TEST raid5f_rebuild_test_sb 00:15:56.110 ************************************ 00:15:56.110 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:56.110 00:15:56.110 real 0m25.422s 00:15:56.110 user 0m32.382s 00:15:56.110 sys 0m3.159s 00:15:56.110 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1128 -- # xtrace_disable 00:15:56.110 11:05:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.110 11:05:01 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:56.110 11:05:01 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:56.110 11:05:01 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:15:56.110 11:05:01 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:15:56.110 11:05:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:56.371 ************************************ 00:15:56.371 START TEST raid_state_function_test_sb_4k 00:15:56.371 ************************************ 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 2 true 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:56.371 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=96419 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 96419' 00:15:56.372 Process raid pid: 96419 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 96419 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@833 -- # '[' -z 96419 ']' 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # local max_retries=100 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:56.372 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # xtrace_disable 00:15:56.372 11:05:01 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.372 [2024-10-29 11:05:01.721069] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:15:56.372 [2024-10-29 11:05:01.721187] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:56.634 [2024-10-29 11:05:01.892883] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:56.634 [2024-10-29 11:05:01.918816] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:56.634 [2024-10-29 11:05:01.962970] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:56.634 [2024-10-29 11:05:01.963012] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@866 -- # return 0 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.205 [2024-10-29 11:05:02.557127] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:57.205 [2024-10-29 11:05:02.557183] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:57.205 [2024-10-29 11:05:02.557199] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:57.205 [2024-10-29 11:05:02.557211] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.205 "name": "Existed_Raid", 00:15:57.205 "uuid": "d2611bc9-0871-48e4-9aae-db68c0fb40c2", 00:15:57.205 "strip_size_kb": 0, 00:15:57.205 "state": "configuring", 00:15:57.205 "raid_level": "raid1", 00:15:57.205 "superblock": true, 00:15:57.205 "num_base_bdevs": 2, 00:15:57.205 "num_base_bdevs_discovered": 0, 00:15:57.205 "num_base_bdevs_operational": 2, 00:15:57.205 "base_bdevs_list": [ 00:15:57.205 { 00:15:57.205 "name": "BaseBdev1", 00:15:57.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.205 "is_configured": false, 00:15:57.205 "data_offset": 0, 00:15:57.205 "data_size": 0 00:15:57.205 }, 00:15:57.205 { 00:15:57.205 "name": "BaseBdev2", 00:15:57.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.205 "is_configured": false, 00:15:57.205 "data_offset": 0, 00:15:57.205 "data_size": 0 00:15:57.205 } 00:15:57.205 ] 00:15:57.205 }' 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.205 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 [2024-10-29 11:05:02.992316] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:57.776 [2024-10-29 11:05:02.992451] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.776 11:05:02 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 [2024-10-29 11:05:03.004318] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:57.776 [2024-10-29 11:05:03.004428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:57.776 [2024-10-29 11:05:03.004476] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:57.776 [2024-10-29 11:05:03.004521] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 [2024-10-29 11:05:03.025571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:57.776 BaseBdev1 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local i 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.776 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.776 [ 00:15:57.776 { 00:15:57.776 "name": "BaseBdev1", 00:15:57.776 "aliases": [ 00:15:57.776 "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19" 00:15:57.776 ], 00:15:57.776 "product_name": "Malloc disk", 00:15:57.777 "block_size": 4096, 00:15:57.777 "num_blocks": 8192, 00:15:57.777 "uuid": "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19", 00:15:57.777 "assigned_rate_limits": { 00:15:57.777 "rw_ios_per_sec": 0, 00:15:57.777 "rw_mbytes_per_sec": 0, 00:15:57.777 "r_mbytes_per_sec": 0, 00:15:57.777 "w_mbytes_per_sec": 0 00:15:57.777 }, 00:15:57.777 "claimed": true, 00:15:57.777 "claim_type": "exclusive_write", 00:15:57.777 "zoned": false, 00:15:57.777 "supported_io_types": { 00:15:57.777 "read": true, 00:15:57.777 "write": true, 00:15:57.777 "unmap": true, 00:15:57.777 "flush": true, 00:15:57.777 "reset": true, 00:15:57.777 "nvme_admin": false, 00:15:57.777 "nvme_io": false, 00:15:57.777 "nvme_io_md": false, 00:15:57.777 "write_zeroes": true, 00:15:57.777 "zcopy": true, 00:15:57.777 "get_zone_info": false, 00:15:57.777 "zone_management": false, 00:15:57.777 "zone_append": false, 00:15:57.777 "compare": false, 00:15:57.777 "compare_and_write": false, 00:15:57.777 "abort": true, 00:15:57.777 "seek_hole": false, 00:15:57.777 "seek_data": false, 00:15:57.777 "copy": true, 00:15:57.777 "nvme_iov_md": false 00:15:57.777 }, 00:15:57.777 "memory_domains": [ 00:15:57.777 { 00:15:57.777 "dma_device_id": "system", 00:15:57.777 "dma_device_type": 1 00:15:57.777 }, 00:15:57.777 { 00:15:57.777 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:57.777 "dma_device_type": 2 00:15:57.777 } 00:15:57.777 ], 00:15:57.777 "driver_specific": {} 00:15:57.777 } 00:15:57.777 ] 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@909 -- # return 0 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.777 "name": "Existed_Raid", 00:15:57.777 "uuid": "a750f0eb-d0e4-4e00-99e2-e60e1386d130", 00:15:57.777 "strip_size_kb": 0, 00:15:57.777 "state": "configuring", 00:15:57.777 "raid_level": "raid1", 00:15:57.777 "superblock": true, 00:15:57.777 "num_base_bdevs": 2, 00:15:57.777 "num_base_bdevs_discovered": 1, 00:15:57.777 "num_base_bdevs_operational": 2, 00:15:57.777 "base_bdevs_list": [ 00:15:57.777 { 00:15:57.777 "name": "BaseBdev1", 00:15:57.777 "uuid": "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19", 00:15:57.777 "is_configured": true, 00:15:57.777 "data_offset": 256, 00:15:57.777 "data_size": 7936 00:15:57.777 }, 00:15:57.777 { 00:15:57.777 "name": "BaseBdev2", 00:15:57.777 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.777 "is_configured": false, 00:15:57.777 "data_offset": 0, 00:15:57.777 "data_size": 0 00:15:57.777 } 00:15:57.777 ] 00:15:57.777 }' 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.777 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.037 [2024-10-29 11:05:03.528750] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:58.037 [2024-10-29 11:05:03.528849] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.037 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.297 [2024-10-29 11:05:03.540753] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:58.297 [2024-10-29 11:05:03.542654] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:58.297 [2024-10-29 11:05:03.542751] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:58.297 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.298 "name": "Existed_Raid", 00:15:58.298 "uuid": "0f5b1213-313d-460c-b24c-93da8efb6315", 00:15:58.298 "strip_size_kb": 0, 00:15:58.298 "state": "configuring", 00:15:58.298 "raid_level": "raid1", 00:15:58.298 "superblock": true, 00:15:58.298 "num_base_bdevs": 2, 00:15:58.298 "num_base_bdevs_discovered": 1, 00:15:58.298 "num_base_bdevs_operational": 2, 00:15:58.298 "base_bdevs_list": [ 00:15:58.298 { 00:15:58.298 "name": "BaseBdev1", 00:15:58.298 "uuid": "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19", 00:15:58.298 "is_configured": true, 00:15:58.298 "data_offset": 256, 00:15:58.298 "data_size": 7936 00:15:58.298 }, 00:15:58.298 { 00:15:58.298 "name": "BaseBdev2", 00:15:58.298 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.298 "is_configured": false, 00:15:58.298 "data_offset": 0, 00:15:58.298 "data_size": 0 00:15:58.298 } 00:15:58.298 ] 00:15:58.298 }' 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.298 11:05:03 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.558 [2024-10-29 11:05:04.027150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:58.558 [2024-10-29 11:05:04.027492] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:15:58.558 [2024-10-29 11:05:04.027551] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:58.558 BaseBdev2 00:15:58.558 [2024-10-29 11:05:04.027908] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:15:58.558 [2024-10-29 11:05:04.028159] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:15:58.558 [2024-10-29 11:05:04.028224] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:15:58.558 [2024-10-29 11:05:04.028453] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local i 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.558 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.818 [ 00:15:58.818 { 00:15:58.818 "name": "BaseBdev2", 00:15:58.818 "aliases": [ 00:15:58.818 "f9336137-04da-4170-af3e-92f106caabb7" 00:15:58.818 ], 00:15:58.818 "product_name": "Malloc disk", 00:15:58.818 "block_size": 4096, 00:15:58.818 "num_blocks": 8192, 00:15:58.818 "uuid": "f9336137-04da-4170-af3e-92f106caabb7", 00:15:58.818 "assigned_rate_limits": { 00:15:58.818 "rw_ios_per_sec": 0, 00:15:58.818 "rw_mbytes_per_sec": 0, 00:15:58.818 "r_mbytes_per_sec": 0, 00:15:58.818 "w_mbytes_per_sec": 0 00:15:58.818 }, 00:15:58.818 "claimed": true, 00:15:58.818 "claim_type": "exclusive_write", 00:15:58.818 "zoned": false, 00:15:58.818 "supported_io_types": { 00:15:58.818 "read": true, 00:15:58.818 "write": true, 00:15:58.818 "unmap": true, 00:15:58.818 "flush": true, 00:15:58.818 "reset": true, 00:15:58.818 "nvme_admin": false, 00:15:58.818 "nvme_io": false, 00:15:58.818 "nvme_io_md": false, 00:15:58.818 "write_zeroes": true, 00:15:58.818 "zcopy": true, 00:15:58.818 "get_zone_info": false, 00:15:58.818 "zone_management": false, 00:15:58.818 "zone_append": false, 00:15:58.818 "compare": false, 00:15:58.818 "compare_and_write": false, 00:15:58.818 "abort": true, 00:15:58.818 "seek_hole": false, 00:15:58.818 "seek_data": false, 00:15:58.818 "copy": true, 00:15:58.818 "nvme_iov_md": false 00:15:58.818 }, 00:15:58.818 "memory_domains": [ 00:15:58.818 { 00:15:58.818 "dma_device_id": "system", 00:15:58.818 "dma_device_type": 1 00:15:58.818 }, 00:15:58.818 { 00:15:58.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:58.818 "dma_device_type": 2 00:15:58.818 } 00:15:58.818 ], 00:15:58.818 "driver_specific": {} 00:15:58.818 } 00:15:58.818 ] 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@909 -- # return 0 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:58.818 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.819 "name": "Existed_Raid", 00:15:58.819 "uuid": "0f5b1213-313d-460c-b24c-93da8efb6315", 00:15:58.819 "strip_size_kb": 0, 00:15:58.819 "state": "online", 00:15:58.819 "raid_level": "raid1", 00:15:58.819 "superblock": true, 00:15:58.819 "num_base_bdevs": 2, 00:15:58.819 "num_base_bdevs_discovered": 2, 00:15:58.819 "num_base_bdevs_operational": 2, 00:15:58.819 "base_bdevs_list": [ 00:15:58.819 { 00:15:58.819 "name": "BaseBdev1", 00:15:58.819 "uuid": "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19", 00:15:58.819 "is_configured": true, 00:15:58.819 "data_offset": 256, 00:15:58.819 "data_size": 7936 00:15:58.819 }, 00:15:58.819 { 00:15:58.819 "name": "BaseBdev2", 00:15:58.819 "uuid": "f9336137-04da-4170-af3e-92f106caabb7", 00:15:58.819 "is_configured": true, 00:15:58.819 "data_offset": 256, 00:15:58.819 "data_size": 7936 00:15:58.819 } 00:15:58.819 ] 00:15:58.819 }' 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.819 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.078 [2024-10-29 11:05:04.510624] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.078 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:59.078 "name": "Existed_Raid", 00:15:59.078 "aliases": [ 00:15:59.078 "0f5b1213-313d-460c-b24c-93da8efb6315" 00:15:59.078 ], 00:15:59.078 "product_name": "Raid Volume", 00:15:59.078 "block_size": 4096, 00:15:59.078 "num_blocks": 7936, 00:15:59.078 "uuid": "0f5b1213-313d-460c-b24c-93da8efb6315", 00:15:59.078 "assigned_rate_limits": { 00:15:59.078 "rw_ios_per_sec": 0, 00:15:59.078 "rw_mbytes_per_sec": 0, 00:15:59.078 "r_mbytes_per_sec": 0, 00:15:59.078 "w_mbytes_per_sec": 0 00:15:59.078 }, 00:15:59.078 "claimed": false, 00:15:59.078 "zoned": false, 00:15:59.078 "supported_io_types": { 00:15:59.078 "read": true, 00:15:59.078 "write": true, 00:15:59.078 "unmap": false, 00:15:59.078 "flush": false, 00:15:59.078 "reset": true, 00:15:59.078 "nvme_admin": false, 00:15:59.078 "nvme_io": false, 00:15:59.078 "nvme_io_md": false, 00:15:59.078 "write_zeroes": true, 00:15:59.078 "zcopy": false, 00:15:59.078 "get_zone_info": false, 00:15:59.078 "zone_management": false, 00:15:59.078 "zone_append": false, 00:15:59.078 "compare": false, 00:15:59.078 "compare_and_write": false, 00:15:59.078 "abort": false, 00:15:59.078 "seek_hole": false, 00:15:59.078 "seek_data": false, 00:15:59.078 "copy": false, 00:15:59.078 "nvme_iov_md": false 00:15:59.078 }, 00:15:59.078 "memory_domains": [ 00:15:59.078 { 00:15:59.078 "dma_device_id": "system", 00:15:59.078 "dma_device_type": 1 00:15:59.078 }, 00:15:59.078 { 00:15:59.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:59.078 "dma_device_type": 2 00:15:59.078 }, 00:15:59.078 { 00:15:59.078 "dma_device_id": "system", 00:15:59.078 "dma_device_type": 1 00:15:59.078 }, 00:15:59.078 { 00:15:59.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:59.078 "dma_device_type": 2 00:15:59.078 } 00:15:59.078 ], 00:15:59.078 "driver_specific": { 00:15:59.078 "raid": { 00:15:59.078 "uuid": "0f5b1213-313d-460c-b24c-93da8efb6315", 00:15:59.078 "strip_size_kb": 0, 00:15:59.078 "state": "online", 00:15:59.078 "raid_level": "raid1", 00:15:59.078 "superblock": true, 00:15:59.078 "num_base_bdevs": 2, 00:15:59.078 "num_base_bdevs_discovered": 2, 00:15:59.078 "num_base_bdevs_operational": 2, 00:15:59.078 "base_bdevs_list": [ 00:15:59.078 { 00:15:59.078 "name": "BaseBdev1", 00:15:59.078 "uuid": "3b6a9fca-ec7d-4a18-a7e8-2851f465cd19", 00:15:59.078 "is_configured": true, 00:15:59.078 "data_offset": 256, 00:15:59.078 "data_size": 7936 00:15:59.078 }, 00:15:59.078 { 00:15:59.079 "name": "BaseBdev2", 00:15:59.079 "uuid": "f9336137-04da-4170-af3e-92f106caabb7", 00:15:59.079 "is_configured": true, 00:15:59.079 "data_offset": 256, 00:15:59.079 "data_size": 7936 00:15:59.079 } 00:15:59.079 ] 00:15:59.079 } 00:15:59.079 } 00:15:59.079 }' 00:15:59.079 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:59.339 BaseBdev2' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.339 [2024-10-29 11:05:04.757972] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.339 "name": "Existed_Raid", 00:15:59.339 "uuid": "0f5b1213-313d-460c-b24c-93da8efb6315", 00:15:59.339 "strip_size_kb": 0, 00:15:59.339 "state": "online", 00:15:59.339 "raid_level": "raid1", 00:15:59.339 "superblock": true, 00:15:59.339 "num_base_bdevs": 2, 00:15:59.339 "num_base_bdevs_discovered": 1, 00:15:59.339 "num_base_bdevs_operational": 1, 00:15:59.339 "base_bdevs_list": [ 00:15:59.339 { 00:15:59.339 "name": null, 00:15:59.339 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.339 "is_configured": false, 00:15:59.339 "data_offset": 0, 00:15:59.339 "data_size": 7936 00:15:59.339 }, 00:15:59.339 { 00:15:59.339 "name": "BaseBdev2", 00:15:59.339 "uuid": "f9336137-04da-4170-af3e-92f106caabb7", 00:15:59.339 "is_configured": true, 00:15:59.339 "data_offset": 256, 00:15:59.339 "data_size": 7936 00:15:59.339 } 00:15:59.339 ] 00:15:59.339 }' 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.339 11:05:04 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:59.912 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.913 [2024-10-29 11:05:05.276463] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:59.913 [2024-10-29 11:05:05.276560] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:59.913 [2024-10-29 11:05:05.288249] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:59.913 [2024-10-29 11:05:05.288421] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:59.913 [2024-10-29 11:05:05.288441] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 96419 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@952 -- # '[' -z 96419 ']' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # kill -0 96419 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@957 -- # uname 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 96419 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:15:59.913 killing process with pid 96419 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@970 -- # echo 'killing process with pid 96419' 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@971 -- # kill 96419 00:15:59.913 [2024-10-29 11:05:05.381900] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:59.913 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@976 -- # wait 96419 00:15:59.913 [2024-10-29 11:05:05.382998] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:00.173 ************************************ 00:16:00.173 END TEST raid_state_function_test_sb_4k 00:16:00.173 ************************************ 00:16:00.173 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:16:00.173 00:16:00.173 real 0m3.981s 00:16:00.173 user 0m6.283s 00:16:00.173 sys 0m0.838s 00:16:00.173 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:00.173 11:05:05 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.173 11:05:05 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:16:00.173 11:05:05 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:16:00.173 11:05:05 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:00.173 11:05:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:00.434 ************************************ 00:16:00.434 START TEST raid_superblock_test_4k 00:16:00.434 ************************************ 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 2 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96663 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96663 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@833 -- # '[' -z 96663 ']' 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:00.434 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:00.434 11:05:05 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.434 [2024-10-29 11:05:05.784428] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:00.434 [2024-10-29 11:05:05.784640] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96663 ] 00:16:00.694 [2024-10-29 11:05:05.958830] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:00.694 [2024-10-29 11:05:05.985472] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:00.694 [2024-10-29 11:05:06.029858] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:00.694 [2024-10-29 11:05:06.029909] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@866 -- # return 0 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 malloc1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 [2024-10-29 11:05:06.609501] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:01.264 [2024-10-29 11:05:06.609664] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.264 [2024-10-29 11:05:06.609709] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:01.264 [2024-10-29 11:05:06.609762] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.264 [2024-10-29 11:05:06.611913] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.264 [2024-10-29 11:05:06.612022] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:01.264 pt1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 malloc2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 [2024-10-29 11:05:06.642246] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:01.264 [2024-10-29 11:05:06.642311] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:01.264 [2024-10-29 11:05:06.642329] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:01.264 [2024-10-29 11:05:06.642342] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:01.264 [2024-10-29 11:05:06.644523] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:01.264 [2024-10-29 11:05:06.644567] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:01.264 pt2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 [2024-10-29 11:05:06.654264] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:01.264 [2024-10-29 11:05:06.656220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:01.264 [2024-10-29 11:05:06.656389] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:01.264 [2024-10-29 11:05:06.656407] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:01.264 [2024-10-29 11:05:06.656680] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:01.264 [2024-10-29 11:05:06.656843] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:01.264 [2024-10-29 11:05:06.656855] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:01.264 [2024-10-29 11:05:06.656989] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.264 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.264 "name": "raid_bdev1", 00:16:01.264 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:01.264 "strip_size_kb": 0, 00:16:01.264 "state": "online", 00:16:01.264 "raid_level": "raid1", 00:16:01.264 "superblock": true, 00:16:01.264 "num_base_bdevs": 2, 00:16:01.264 "num_base_bdevs_discovered": 2, 00:16:01.264 "num_base_bdevs_operational": 2, 00:16:01.264 "base_bdevs_list": [ 00:16:01.264 { 00:16:01.264 "name": "pt1", 00:16:01.264 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:01.264 "is_configured": true, 00:16:01.264 "data_offset": 256, 00:16:01.264 "data_size": 7936 00:16:01.264 }, 00:16:01.264 { 00:16:01.264 "name": "pt2", 00:16:01.264 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.265 "is_configured": true, 00:16:01.265 "data_offset": 256, 00:16:01.265 "data_size": 7936 00:16:01.265 } 00:16:01.265 ] 00:16:01.265 }' 00:16:01.265 11:05:06 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.265 11:05:06 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.835 [2024-10-29 11:05:07.125696] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:01.835 "name": "raid_bdev1", 00:16:01.835 "aliases": [ 00:16:01.835 "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc" 00:16:01.835 ], 00:16:01.835 "product_name": "Raid Volume", 00:16:01.835 "block_size": 4096, 00:16:01.835 "num_blocks": 7936, 00:16:01.835 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:01.835 "assigned_rate_limits": { 00:16:01.835 "rw_ios_per_sec": 0, 00:16:01.835 "rw_mbytes_per_sec": 0, 00:16:01.835 "r_mbytes_per_sec": 0, 00:16:01.835 "w_mbytes_per_sec": 0 00:16:01.835 }, 00:16:01.835 "claimed": false, 00:16:01.835 "zoned": false, 00:16:01.835 "supported_io_types": { 00:16:01.835 "read": true, 00:16:01.835 "write": true, 00:16:01.835 "unmap": false, 00:16:01.835 "flush": false, 00:16:01.835 "reset": true, 00:16:01.835 "nvme_admin": false, 00:16:01.835 "nvme_io": false, 00:16:01.835 "nvme_io_md": false, 00:16:01.835 "write_zeroes": true, 00:16:01.835 "zcopy": false, 00:16:01.835 "get_zone_info": false, 00:16:01.835 "zone_management": false, 00:16:01.835 "zone_append": false, 00:16:01.835 "compare": false, 00:16:01.835 "compare_and_write": false, 00:16:01.835 "abort": false, 00:16:01.835 "seek_hole": false, 00:16:01.835 "seek_data": false, 00:16:01.835 "copy": false, 00:16:01.835 "nvme_iov_md": false 00:16:01.835 }, 00:16:01.835 "memory_domains": [ 00:16:01.835 { 00:16:01.835 "dma_device_id": "system", 00:16:01.835 "dma_device_type": 1 00:16:01.835 }, 00:16:01.835 { 00:16:01.835 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.835 "dma_device_type": 2 00:16:01.835 }, 00:16:01.835 { 00:16:01.835 "dma_device_id": "system", 00:16:01.835 "dma_device_type": 1 00:16:01.835 }, 00:16:01.835 { 00:16:01.835 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.835 "dma_device_type": 2 00:16:01.835 } 00:16:01.835 ], 00:16:01.835 "driver_specific": { 00:16:01.835 "raid": { 00:16:01.835 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:01.835 "strip_size_kb": 0, 00:16:01.835 "state": "online", 00:16:01.835 "raid_level": "raid1", 00:16:01.835 "superblock": true, 00:16:01.835 "num_base_bdevs": 2, 00:16:01.835 "num_base_bdevs_discovered": 2, 00:16:01.835 "num_base_bdevs_operational": 2, 00:16:01.835 "base_bdevs_list": [ 00:16:01.835 { 00:16:01.835 "name": "pt1", 00:16:01.835 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:01.835 "is_configured": true, 00:16:01.835 "data_offset": 256, 00:16:01.835 "data_size": 7936 00:16:01.835 }, 00:16:01.835 { 00:16:01.835 "name": "pt2", 00:16:01.835 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:01.835 "is_configured": true, 00:16:01.835 "data_offset": 256, 00:16:01.835 "data_size": 7936 00:16:01.835 } 00:16:01.835 ] 00:16:01.835 } 00:16:01.835 } 00:16:01.835 }' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:01.835 pt2' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:01.835 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.836 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.096 [2024-10-29 11:05:07.381172] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=97ff0bda-012b-4fe6-8e95-eeba0cbf3abc 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z 97ff0bda-012b-4fe6-8e95-eeba0cbf3abc ']' 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.096 [2024-10-29 11:05:07.424894] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.096 [2024-10-29 11:05:07.424922] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:02.096 [2024-10-29 11:05:07.425002] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:02.096 [2024-10-29 11:05:07.425076] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:02.096 [2024-10-29 11:05:07.425088] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.096 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # local es=0 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.097 [2024-10-29 11:05:07.560693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:02.097 [2024-10-29 11:05:07.562556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:02.097 [2024-10-29 11:05:07.562673] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:02.097 [2024-10-29 11:05:07.562784] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:02.097 [2024-10-29 11:05:07.562864] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:02.097 [2024-10-29 11:05:07.562903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:16:02.097 request: 00:16:02.097 { 00:16:02.097 "name": "raid_bdev1", 00:16:02.097 "raid_level": "raid1", 00:16:02.097 "base_bdevs": [ 00:16:02.097 "malloc1", 00:16:02.097 "malloc2" 00:16:02.097 ], 00:16:02.097 "superblock": false, 00:16:02.097 "method": "bdev_raid_create", 00:16:02.097 "req_id": 1 00:16:02.097 } 00:16:02.097 Got JSON-RPC error response 00:16:02.097 response: 00:16:02.097 { 00:16:02.097 "code": -17, 00:16:02.097 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:02.097 } 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # es=1 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.097 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.357 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:02.357 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:02.357 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:02.357 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.357 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.357 [2024-10-29 11:05:07.624516] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:02.357 [2024-10-29 11:05:07.624567] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.357 [2024-10-29 11:05:07.624586] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:02.357 [2024-10-29 11:05:07.624596] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.358 [2024-10-29 11:05:07.626619] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.358 [2024-10-29 11:05:07.626659] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:02.358 [2024-10-29 11:05:07.626732] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:02.358 [2024-10-29 11:05:07.626781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:02.358 pt1 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.358 "name": "raid_bdev1", 00:16:02.358 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:02.358 "strip_size_kb": 0, 00:16:02.358 "state": "configuring", 00:16:02.358 "raid_level": "raid1", 00:16:02.358 "superblock": true, 00:16:02.358 "num_base_bdevs": 2, 00:16:02.358 "num_base_bdevs_discovered": 1, 00:16:02.358 "num_base_bdevs_operational": 2, 00:16:02.358 "base_bdevs_list": [ 00:16:02.358 { 00:16:02.358 "name": "pt1", 00:16:02.358 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:02.358 "is_configured": true, 00:16:02.358 "data_offset": 256, 00:16:02.358 "data_size": 7936 00:16:02.358 }, 00:16:02.358 { 00:16:02.358 "name": null, 00:16:02.358 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.358 "is_configured": false, 00:16:02.358 "data_offset": 256, 00:16:02.358 "data_size": 7936 00:16:02.358 } 00:16:02.358 ] 00:16:02.358 }' 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.358 11:05:07 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.618 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.619 [2024-10-29 11:05:08.072039] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:02.619 [2024-10-29 11:05:08.072169] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:02.619 [2024-10-29 11:05:08.072212] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:02.619 [2024-10-29 11:05:08.072247] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:02.619 [2024-10-29 11:05:08.072680] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:02.619 [2024-10-29 11:05:08.072746] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:02.619 [2024-10-29 11:05:08.072854] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:02.619 [2024-10-29 11:05:08.072925] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:02.619 [2024-10-29 11:05:08.073068] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:02.619 [2024-10-29 11:05:08.073112] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:02.619 [2024-10-29 11:05:08.073387] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:02.619 [2024-10-29 11:05:08.073564] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:02.619 [2024-10-29 11:05:08.073618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:16:02.619 [2024-10-29 11:05:08.073827] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:02.619 pt2 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.619 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.880 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.880 "name": "raid_bdev1", 00:16:02.880 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:02.880 "strip_size_kb": 0, 00:16:02.880 "state": "online", 00:16:02.880 "raid_level": "raid1", 00:16:02.880 "superblock": true, 00:16:02.880 "num_base_bdevs": 2, 00:16:02.880 "num_base_bdevs_discovered": 2, 00:16:02.880 "num_base_bdevs_operational": 2, 00:16:02.880 "base_bdevs_list": [ 00:16:02.880 { 00:16:02.880 "name": "pt1", 00:16:02.880 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:02.880 "is_configured": true, 00:16:02.880 "data_offset": 256, 00:16:02.880 "data_size": 7936 00:16:02.880 }, 00:16:02.880 { 00:16:02.880 "name": "pt2", 00:16:02.880 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:02.880 "is_configured": true, 00:16:02.880 "data_offset": 256, 00:16:02.880 "data_size": 7936 00:16:02.880 } 00:16:02.880 ] 00:16:02.880 }' 00:16:02.880 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.880 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.141 [2024-10-29 11:05:08.551468] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:03.141 "name": "raid_bdev1", 00:16:03.141 "aliases": [ 00:16:03.141 "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc" 00:16:03.141 ], 00:16:03.141 "product_name": "Raid Volume", 00:16:03.141 "block_size": 4096, 00:16:03.141 "num_blocks": 7936, 00:16:03.141 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:03.141 "assigned_rate_limits": { 00:16:03.141 "rw_ios_per_sec": 0, 00:16:03.141 "rw_mbytes_per_sec": 0, 00:16:03.141 "r_mbytes_per_sec": 0, 00:16:03.141 "w_mbytes_per_sec": 0 00:16:03.141 }, 00:16:03.141 "claimed": false, 00:16:03.141 "zoned": false, 00:16:03.141 "supported_io_types": { 00:16:03.141 "read": true, 00:16:03.141 "write": true, 00:16:03.141 "unmap": false, 00:16:03.141 "flush": false, 00:16:03.141 "reset": true, 00:16:03.141 "nvme_admin": false, 00:16:03.141 "nvme_io": false, 00:16:03.141 "nvme_io_md": false, 00:16:03.141 "write_zeroes": true, 00:16:03.141 "zcopy": false, 00:16:03.141 "get_zone_info": false, 00:16:03.141 "zone_management": false, 00:16:03.141 "zone_append": false, 00:16:03.141 "compare": false, 00:16:03.141 "compare_and_write": false, 00:16:03.141 "abort": false, 00:16:03.141 "seek_hole": false, 00:16:03.141 "seek_data": false, 00:16:03.141 "copy": false, 00:16:03.141 "nvme_iov_md": false 00:16:03.141 }, 00:16:03.141 "memory_domains": [ 00:16:03.141 { 00:16:03.141 "dma_device_id": "system", 00:16:03.141 "dma_device_type": 1 00:16:03.141 }, 00:16:03.141 { 00:16:03.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:03.141 "dma_device_type": 2 00:16:03.141 }, 00:16:03.141 { 00:16:03.141 "dma_device_id": "system", 00:16:03.141 "dma_device_type": 1 00:16:03.141 }, 00:16:03.141 { 00:16:03.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:03.141 "dma_device_type": 2 00:16:03.141 } 00:16:03.141 ], 00:16:03.141 "driver_specific": { 00:16:03.141 "raid": { 00:16:03.141 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:03.141 "strip_size_kb": 0, 00:16:03.141 "state": "online", 00:16:03.141 "raid_level": "raid1", 00:16:03.141 "superblock": true, 00:16:03.141 "num_base_bdevs": 2, 00:16:03.141 "num_base_bdevs_discovered": 2, 00:16:03.141 "num_base_bdevs_operational": 2, 00:16:03.141 "base_bdevs_list": [ 00:16:03.141 { 00:16:03.141 "name": "pt1", 00:16:03.141 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:03.141 "is_configured": true, 00:16:03.141 "data_offset": 256, 00:16:03.141 "data_size": 7936 00:16:03.141 }, 00:16:03.141 { 00:16:03.141 "name": "pt2", 00:16:03.141 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:03.141 "is_configured": true, 00:16:03.141 "data_offset": 256, 00:16:03.141 "data_size": 7936 00:16:03.141 } 00:16:03.141 ] 00:16:03.141 } 00:16:03.141 } 00:16:03.141 }' 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:03.141 pt2' 00:16:03.141 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:03.402 [2024-10-29 11:05:08.779027] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' 97ff0bda-012b-4fe6-8e95-eeba0cbf3abc '!=' 97ff0bda-012b-4fe6-8e95-eeba0cbf3abc ']' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.402 [2024-10-29 11:05:08.826743] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:03.402 "name": "raid_bdev1", 00:16:03.402 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:03.402 "strip_size_kb": 0, 00:16:03.402 "state": "online", 00:16:03.402 "raid_level": "raid1", 00:16:03.402 "superblock": true, 00:16:03.402 "num_base_bdevs": 2, 00:16:03.402 "num_base_bdevs_discovered": 1, 00:16:03.402 "num_base_bdevs_operational": 1, 00:16:03.402 "base_bdevs_list": [ 00:16:03.402 { 00:16:03.402 "name": null, 00:16:03.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:03.402 "is_configured": false, 00:16:03.402 "data_offset": 0, 00:16:03.402 "data_size": 7936 00:16:03.402 }, 00:16:03.402 { 00:16:03.402 "name": "pt2", 00:16:03.402 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:03.402 "is_configured": true, 00:16:03.402 "data_offset": 256, 00:16:03.402 "data_size": 7936 00:16:03.402 } 00:16:03.402 ] 00:16:03.402 }' 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:03.402 11:05:08 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 [2024-10-29 11:05:09.285986] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:03.973 [2024-10-29 11:05:09.286071] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:03.973 [2024-10-29 11:05:09.286165] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:03.973 [2024-10-29 11:05:09.286254] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:03.973 [2024-10-29 11:05:09.286326] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 [2024-10-29 11:05:09.357872] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:03.973 [2024-10-29 11:05:09.357972] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:03.973 [2024-10-29 11:05:09.358010] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:03.973 [2024-10-29 11:05:09.358041] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:03.973 [2024-10-29 11:05:09.360235] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:03.973 [2024-10-29 11:05:09.360315] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:03.973 [2024-10-29 11:05:09.360437] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:03.973 [2024-10-29 11:05:09.360498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:03.973 [2024-10-29 11:05:09.360629] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:16:03.973 [2024-10-29 11:05:09.360673] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:03.973 [2024-10-29 11:05:09.360934] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:03.973 [2024-10-29 11:05:09.361113] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:16:03.973 [2024-10-29 11:05:09.361164] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:16:03.973 [2024-10-29 11:05:09.361335] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:03.973 pt2 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:03.973 "name": "raid_bdev1", 00:16:03.973 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:03.973 "strip_size_kb": 0, 00:16:03.973 "state": "online", 00:16:03.973 "raid_level": "raid1", 00:16:03.973 "superblock": true, 00:16:03.973 "num_base_bdevs": 2, 00:16:03.973 "num_base_bdevs_discovered": 1, 00:16:03.973 "num_base_bdevs_operational": 1, 00:16:03.973 "base_bdevs_list": [ 00:16:03.973 { 00:16:03.973 "name": null, 00:16:03.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:03.973 "is_configured": false, 00:16:03.973 "data_offset": 256, 00:16:03.973 "data_size": 7936 00:16:03.973 }, 00:16:03.973 { 00:16:03.973 "name": "pt2", 00:16:03.973 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:03.973 "is_configured": true, 00:16:03.973 "data_offset": 256, 00:16:03.973 "data_size": 7936 00:16:03.973 } 00:16:03.973 ] 00:16:03.973 }' 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:03.973 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.544 [2024-10-29 11:05:09.833081] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:04.544 [2024-10-29 11:05:09.833167] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:04.544 [2024-10-29 11:05:09.833229] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:04.544 [2024-10-29 11:05:09.833269] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:04.544 [2024-10-29 11:05:09.833282] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.544 [2024-10-29 11:05:09.896933] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:04.544 [2024-10-29 11:05:09.896992] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.544 [2024-10-29 11:05:09.897016] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:04.544 [2024-10-29 11:05:09.897032] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.544 [2024-10-29 11:05:09.899131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.544 [2024-10-29 11:05:09.899234] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:04.544 [2024-10-29 11:05:09.899303] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:04.544 [2024-10-29 11:05:09.899343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:04.544 [2024-10-29 11:05:09.899507] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:04.544 [2024-10-29 11:05:09.899533] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:04.544 [2024-10-29 11:05:09.899553] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:16:04.544 [2024-10-29 11:05:09.899596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:04.544 [2024-10-29 11:05:09.899669] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:16:04.544 [2024-10-29 11:05:09.899682] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:04.544 [2024-10-29 11:05:09.899896] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:04.544 [2024-10-29 11:05:09.900010] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:16:04.544 [2024-10-29 11:05:09.900031] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:16:04.544 [2024-10-29 11:05:09.900142] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.544 pt1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.544 "name": "raid_bdev1", 00:16:04.544 "uuid": "97ff0bda-012b-4fe6-8e95-eeba0cbf3abc", 00:16:04.544 "strip_size_kb": 0, 00:16:04.544 "state": "online", 00:16:04.544 "raid_level": "raid1", 00:16:04.544 "superblock": true, 00:16:04.544 "num_base_bdevs": 2, 00:16:04.544 "num_base_bdevs_discovered": 1, 00:16:04.544 "num_base_bdevs_operational": 1, 00:16:04.544 "base_bdevs_list": [ 00:16:04.544 { 00:16:04.544 "name": null, 00:16:04.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:04.544 "is_configured": false, 00:16:04.544 "data_offset": 256, 00:16:04.544 "data_size": 7936 00:16:04.544 }, 00:16:04.544 { 00:16:04.544 "name": "pt2", 00:16:04.544 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:04.544 "is_configured": true, 00:16:04.544 "data_offset": 256, 00:16:04.544 "data_size": 7936 00:16:04.544 } 00:16:04.544 ] 00:16:04.544 }' 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.544 11:05:09 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.116 [2024-10-29 11:05:10.456209] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' 97ff0bda-012b-4fe6-8e95-eeba0cbf3abc '!=' 97ff0bda-012b-4fe6-8e95-eeba0cbf3abc ']' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96663 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@952 -- # '[' -z 96663 ']' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # kill -0 96663 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@957 -- # uname 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 96663 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@970 -- # echo 'killing process with pid 96663' 00:16:05.116 killing process with pid 96663 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@971 -- # kill 96663 00:16:05.116 [2024-10-29 11:05:10.535433] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:05.116 [2024-10-29 11:05:10.535502] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:05.116 [2024-10-29 11:05:10.535543] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:05.116 [2024-10-29 11:05:10.535552] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:16:05.116 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@976 -- # wait 96663 00:16:05.116 [2024-10-29 11:05:10.557975] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:05.376 11:05:10 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:16:05.376 00:16:05.376 real 0m5.093s 00:16:05.376 user 0m8.325s 00:16:05.376 sys 0m1.144s 00:16:05.376 ************************************ 00:16:05.376 END TEST raid_superblock_test_4k 00:16:05.376 ************************************ 00:16:05.376 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:05.377 11:05:10 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.377 11:05:10 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:16:05.377 11:05:10 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:16:05.377 11:05:10 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:16:05.377 11:05:10 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:05.377 11:05:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:05.377 ************************************ 00:16:05.377 START TEST raid_rebuild_test_sb_4k 00:16:05.377 ************************************ 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 true false true 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96980 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96980 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@833 -- # '[' -z 96980 ']' 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:05.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:05.377 11:05:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.637 [2024-10-29 11:05:10.948523] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:05.637 [2024-10-29 11:05:10.948761] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:16:05.637 Zero copy mechanism will not be used. 00:16:05.637 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96980 ] 00:16:05.637 [2024-10-29 11:05:11.117895] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:05.897 [2024-10-29 11:05:11.144663] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:05.897 [2024-10-29 11:05:11.188860] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:05.897 [2024-10-29 11:05:11.188918] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@866 -- # return 0 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 BaseBdev1_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 [2024-10-29 11:05:11.768126] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:06.468 [2024-10-29 11:05:11.768216] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.468 [2024-10-29 11:05:11.768243] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:06.468 [2024-10-29 11:05:11.768266] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.468 [2024-10-29 11:05:11.770369] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.468 [2024-10-29 11:05:11.770418] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:06.468 BaseBdev1 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 BaseBdev2_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 [2024-10-29 11:05:11.792914] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:06.468 [2024-10-29 11:05:11.792976] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.468 [2024-10-29 11:05:11.792998] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:06.468 [2024-10-29 11:05:11.793009] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.468 [2024-10-29 11:05:11.795080] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.468 [2024-10-29 11:05:11.795131] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:06.468 BaseBdev2 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 spare_malloc 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 spare_delay 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.468 [2024-10-29 11:05:11.833699] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:06.468 [2024-10-29 11:05:11.833760] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.468 [2024-10-29 11:05:11.833782] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:06.468 [2024-10-29 11:05:11.833793] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.468 [2024-10-29 11:05:11.835850] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.468 [2024-10-29 11:05:11.835904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:06.468 spare 00:16:06.468 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.469 [2024-10-29 11:05:11.845737] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:06.469 [2024-10-29 11:05:11.847574] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:06.469 [2024-10-29 11:05:11.847746] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:06.469 [2024-10-29 11:05:11.847760] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:06.469 [2024-10-29 11:05:11.848008] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:06.469 [2024-10-29 11:05:11.848156] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:06.469 [2024-10-29 11:05:11.848181] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:06.469 [2024-10-29 11:05:11.848304] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.469 "name": "raid_bdev1", 00:16:06.469 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:06.469 "strip_size_kb": 0, 00:16:06.469 "state": "online", 00:16:06.469 "raid_level": "raid1", 00:16:06.469 "superblock": true, 00:16:06.469 "num_base_bdevs": 2, 00:16:06.469 "num_base_bdevs_discovered": 2, 00:16:06.469 "num_base_bdevs_operational": 2, 00:16:06.469 "base_bdevs_list": [ 00:16:06.469 { 00:16:06.469 "name": "BaseBdev1", 00:16:06.469 "uuid": "4b7a4b39-1e35-577e-b72d-afd0cda629a8", 00:16:06.469 "is_configured": true, 00:16:06.469 "data_offset": 256, 00:16:06.469 "data_size": 7936 00:16:06.469 }, 00:16:06.469 { 00:16:06.469 "name": "BaseBdev2", 00:16:06.469 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:06.469 "is_configured": true, 00:16:06.469 "data_offset": 256, 00:16:06.469 "data_size": 7936 00:16:06.469 } 00:16:06.469 ] 00:16:06.469 }' 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.469 11:05:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.040 [2024-10-29 11:05:12.297140] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:07.040 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:07.301 [2024-10-29 11:05:12.564525] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:07.301 /dev/nbd0 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # local i 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # break 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:07.301 1+0 records in 00:16:07.301 1+0 records out 00:16:07.301 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000541498 s, 7.6 MB/s 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # size=4096 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # return 0 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:07.301 11:05:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:07.871 7936+0 records in 00:16:07.871 7936+0 records out 00:16:07.871 32505856 bytes (33 MB, 31 MiB) copied, 0.663047 s, 49.0 MB/s 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:07.871 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:08.131 [2024-10-29 11:05:13.518464] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.131 [2024-10-29 11:05:13.531447] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.131 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.131 "name": "raid_bdev1", 00:16:08.131 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:08.131 "strip_size_kb": 0, 00:16:08.131 "state": "online", 00:16:08.131 "raid_level": "raid1", 00:16:08.131 "superblock": true, 00:16:08.131 "num_base_bdevs": 2, 00:16:08.131 "num_base_bdevs_discovered": 1, 00:16:08.131 "num_base_bdevs_operational": 1, 00:16:08.131 "base_bdevs_list": [ 00:16:08.131 { 00:16:08.131 "name": null, 00:16:08.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.132 "is_configured": false, 00:16:08.132 "data_offset": 0, 00:16:08.132 "data_size": 7936 00:16:08.132 }, 00:16:08.132 { 00:16:08.132 "name": "BaseBdev2", 00:16:08.132 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:08.132 "is_configured": true, 00:16:08.132 "data_offset": 256, 00:16:08.132 "data_size": 7936 00:16:08.132 } 00:16:08.132 ] 00:16:08.132 }' 00:16:08.132 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.132 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.701 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:08.701 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.701 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.701 [2024-10-29 11:05:13.978646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:08.702 [2024-10-29 11:05:13.993594] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d0c0 00:16:08.702 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.702 11:05:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:08.702 [2024-10-29 11:05:13.995918] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:09.642 11:05:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:09.642 "name": "raid_bdev1", 00:16:09.642 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:09.642 "strip_size_kb": 0, 00:16:09.642 "state": "online", 00:16:09.642 "raid_level": "raid1", 00:16:09.642 "superblock": true, 00:16:09.642 "num_base_bdevs": 2, 00:16:09.642 "num_base_bdevs_discovered": 2, 00:16:09.642 "num_base_bdevs_operational": 2, 00:16:09.642 "process": { 00:16:09.642 "type": "rebuild", 00:16:09.642 "target": "spare", 00:16:09.642 "progress": { 00:16:09.642 "blocks": 2560, 00:16:09.642 "percent": 32 00:16:09.642 } 00:16:09.642 }, 00:16:09.642 "base_bdevs_list": [ 00:16:09.642 { 00:16:09.642 "name": "spare", 00:16:09.642 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:09.642 "is_configured": true, 00:16:09.642 "data_offset": 256, 00:16:09.642 "data_size": 7936 00:16:09.642 }, 00:16:09.642 { 00:16:09.642 "name": "BaseBdev2", 00:16:09.642 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:09.642 "is_configured": true, 00:16:09.642 "data_offset": 256, 00:16:09.642 "data_size": 7936 00:16:09.642 } 00:16:09.642 ] 00:16:09.642 }' 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.642 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.902 [2024-10-29 11:05:15.148319] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:09.902 [2024-10-29 11:05:15.200735] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:09.902 [2024-10-29 11:05:15.200823] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:09.902 [2024-10-29 11:05:15.200848] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:09.902 [2024-10-29 11:05:15.200857] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.902 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.902 "name": "raid_bdev1", 00:16:09.902 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:09.902 "strip_size_kb": 0, 00:16:09.902 "state": "online", 00:16:09.902 "raid_level": "raid1", 00:16:09.902 "superblock": true, 00:16:09.902 "num_base_bdevs": 2, 00:16:09.902 "num_base_bdevs_discovered": 1, 00:16:09.902 "num_base_bdevs_operational": 1, 00:16:09.902 "base_bdevs_list": [ 00:16:09.902 { 00:16:09.902 "name": null, 00:16:09.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:09.902 "is_configured": false, 00:16:09.902 "data_offset": 0, 00:16:09.902 "data_size": 7936 00:16:09.902 }, 00:16:09.902 { 00:16:09.902 "name": "BaseBdev2", 00:16:09.902 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:09.903 "is_configured": true, 00:16:09.903 "data_offset": 256, 00:16:09.903 "data_size": 7936 00:16:09.903 } 00:16:09.903 ] 00:16:09.903 }' 00:16:09.903 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.903 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.162 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:10.162 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.163 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:10.163 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:10.163 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.163 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.163 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.423 "name": "raid_bdev1", 00:16:10.423 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:10.423 "strip_size_kb": 0, 00:16:10.423 "state": "online", 00:16:10.423 "raid_level": "raid1", 00:16:10.423 "superblock": true, 00:16:10.423 "num_base_bdevs": 2, 00:16:10.423 "num_base_bdevs_discovered": 1, 00:16:10.423 "num_base_bdevs_operational": 1, 00:16:10.423 "base_bdevs_list": [ 00:16:10.423 { 00:16:10.423 "name": null, 00:16:10.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.423 "is_configured": false, 00:16:10.423 "data_offset": 0, 00:16:10.423 "data_size": 7936 00:16:10.423 }, 00:16:10.423 { 00:16:10.423 "name": "BaseBdev2", 00:16:10.423 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:10.423 "is_configured": true, 00:16:10.423 "data_offset": 256, 00:16:10.423 "data_size": 7936 00:16:10.423 } 00:16:10.423 ] 00:16:10.423 }' 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.423 [2024-10-29 11:05:15.768634] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:10.423 [2024-10-29 11:05:15.773441] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d190 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.423 11:05:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:10.423 [2024-10-29 11:05:15.775332] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:11.364 "name": "raid_bdev1", 00:16:11.364 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:11.364 "strip_size_kb": 0, 00:16:11.364 "state": "online", 00:16:11.364 "raid_level": "raid1", 00:16:11.364 "superblock": true, 00:16:11.364 "num_base_bdevs": 2, 00:16:11.364 "num_base_bdevs_discovered": 2, 00:16:11.364 "num_base_bdevs_operational": 2, 00:16:11.364 "process": { 00:16:11.364 "type": "rebuild", 00:16:11.364 "target": "spare", 00:16:11.364 "progress": { 00:16:11.364 "blocks": 2560, 00:16:11.364 "percent": 32 00:16:11.364 } 00:16:11.364 }, 00:16:11.364 "base_bdevs_list": [ 00:16:11.364 { 00:16:11.364 "name": "spare", 00:16:11.364 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:11.364 "is_configured": true, 00:16:11.364 "data_offset": 256, 00:16:11.364 "data_size": 7936 00:16:11.364 }, 00:16:11.364 { 00:16:11.364 "name": "BaseBdev2", 00:16:11.364 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:11.364 "is_configured": true, 00:16:11.364 "data_offset": 256, 00:16:11.364 "data_size": 7936 00:16:11.364 } 00:16:11.364 ] 00:16:11.364 }' 00:16:11.364 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:11.624 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=561 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.624 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:11.624 "name": "raid_bdev1", 00:16:11.624 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:11.624 "strip_size_kb": 0, 00:16:11.624 "state": "online", 00:16:11.624 "raid_level": "raid1", 00:16:11.624 "superblock": true, 00:16:11.624 "num_base_bdevs": 2, 00:16:11.624 "num_base_bdevs_discovered": 2, 00:16:11.624 "num_base_bdevs_operational": 2, 00:16:11.624 "process": { 00:16:11.624 "type": "rebuild", 00:16:11.624 "target": "spare", 00:16:11.624 "progress": { 00:16:11.624 "blocks": 2816, 00:16:11.625 "percent": 35 00:16:11.625 } 00:16:11.625 }, 00:16:11.625 "base_bdevs_list": [ 00:16:11.625 { 00:16:11.625 "name": "spare", 00:16:11.625 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:11.625 "is_configured": true, 00:16:11.625 "data_offset": 256, 00:16:11.625 "data_size": 7936 00:16:11.625 }, 00:16:11.625 { 00:16:11.625 "name": "BaseBdev2", 00:16:11.625 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:11.625 "is_configured": true, 00:16:11.625 "data_offset": 256, 00:16:11.625 "data_size": 7936 00:16:11.625 } 00:16:11.625 ] 00:16:11.625 }' 00:16:11.625 11:05:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:11.625 11:05:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:11.625 11:05:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:11.625 11:05:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:11.625 11:05:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.593 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:12.864 "name": "raid_bdev1", 00:16:12.864 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:12.864 "strip_size_kb": 0, 00:16:12.864 "state": "online", 00:16:12.864 "raid_level": "raid1", 00:16:12.864 "superblock": true, 00:16:12.864 "num_base_bdevs": 2, 00:16:12.864 "num_base_bdevs_discovered": 2, 00:16:12.864 "num_base_bdevs_operational": 2, 00:16:12.864 "process": { 00:16:12.864 "type": "rebuild", 00:16:12.864 "target": "spare", 00:16:12.864 "progress": { 00:16:12.864 "blocks": 5888, 00:16:12.864 "percent": 74 00:16:12.864 } 00:16:12.864 }, 00:16:12.864 "base_bdevs_list": [ 00:16:12.864 { 00:16:12.864 "name": "spare", 00:16:12.864 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:12.864 "is_configured": true, 00:16:12.864 "data_offset": 256, 00:16:12.864 "data_size": 7936 00:16:12.864 }, 00:16:12.864 { 00:16:12.864 "name": "BaseBdev2", 00:16:12.864 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:12.864 "is_configured": true, 00:16:12.864 "data_offset": 256, 00:16:12.864 "data_size": 7936 00:16:12.864 } 00:16:12.864 ] 00:16:12.864 }' 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:12.864 11:05:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:13.435 [2024-10-29 11:05:18.885809] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:13.435 [2024-10-29 11:05:18.885888] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:13.435 [2024-10-29 11:05:18.886002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.006 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.006 "name": "raid_bdev1", 00:16:14.006 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:14.006 "strip_size_kb": 0, 00:16:14.006 "state": "online", 00:16:14.006 "raid_level": "raid1", 00:16:14.006 "superblock": true, 00:16:14.006 "num_base_bdevs": 2, 00:16:14.006 "num_base_bdevs_discovered": 2, 00:16:14.006 "num_base_bdevs_operational": 2, 00:16:14.006 "base_bdevs_list": [ 00:16:14.006 { 00:16:14.006 "name": "spare", 00:16:14.006 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:14.006 "is_configured": true, 00:16:14.006 "data_offset": 256, 00:16:14.006 "data_size": 7936 00:16:14.006 }, 00:16:14.006 { 00:16:14.006 "name": "BaseBdev2", 00:16:14.006 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:14.006 "is_configured": true, 00:16:14.006 "data_offset": 256, 00:16:14.006 "data_size": 7936 00:16:14.006 } 00:16:14.006 ] 00:16:14.006 }' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.007 "name": "raid_bdev1", 00:16:14.007 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:14.007 "strip_size_kb": 0, 00:16:14.007 "state": "online", 00:16:14.007 "raid_level": "raid1", 00:16:14.007 "superblock": true, 00:16:14.007 "num_base_bdevs": 2, 00:16:14.007 "num_base_bdevs_discovered": 2, 00:16:14.007 "num_base_bdevs_operational": 2, 00:16:14.007 "base_bdevs_list": [ 00:16:14.007 { 00:16:14.007 "name": "spare", 00:16:14.007 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:14.007 "is_configured": true, 00:16:14.007 "data_offset": 256, 00:16:14.007 "data_size": 7936 00:16:14.007 }, 00:16:14.007 { 00:16:14.007 "name": "BaseBdev2", 00:16:14.007 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:14.007 "is_configured": true, 00:16:14.007 "data_offset": 256, 00:16:14.007 "data_size": 7936 00:16:14.007 } 00:16:14.007 ] 00:16:14.007 }' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.007 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.267 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.267 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:14.267 "name": "raid_bdev1", 00:16:14.267 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:14.267 "strip_size_kb": 0, 00:16:14.267 "state": "online", 00:16:14.267 "raid_level": "raid1", 00:16:14.267 "superblock": true, 00:16:14.267 "num_base_bdevs": 2, 00:16:14.267 "num_base_bdevs_discovered": 2, 00:16:14.267 "num_base_bdevs_operational": 2, 00:16:14.267 "base_bdevs_list": [ 00:16:14.267 { 00:16:14.267 "name": "spare", 00:16:14.267 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:14.267 "is_configured": true, 00:16:14.267 "data_offset": 256, 00:16:14.267 "data_size": 7936 00:16:14.267 }, 00:16:14.267 { 00:16:14.267 "name": "BaseBdev2", 00:16:14.267 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:14.267 "is_configured": true, 00:16:14.267 "data_offset": 256, 00:16:14.267 "data_size": 7936 00:16:14.267 } 00:16:14.267 ] 00:16:14.267 }' 00:16:14.267 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:14.267 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.528 [2024-10-29 11:05:19.956765] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:14.528 [2024-10-29 11:05:19.956857] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:14.528 [2024-10-29 11:05:19.956960] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:14.528 [2024-10-29 11:05:19.957099] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:14.528 [2024-10-29 11:05:19.957157] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:16:14.528 11:05:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:14.528 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:14.787 /dev/nbd0 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # local i 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # break 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:14.788 1+0 records in 00:16:14.788 1+0 records out 00:16:14.788 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00056689 s, 7.2 MB/s 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # size=4096 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # return 0 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:14.788 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:15.048 /dev/nbd1 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # local i 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # break 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:15.048 1+0 records in 00:16:15.048 1+0 records out 00:16:15.048 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000419516 s, 9.8 MB/s 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # size=4096 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # return 0 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:15.048 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:15.309 11:05:20 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.570 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:15.570 [2024-10-29 11:05:21.035890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:15.570 [2024-10-29 11:05:21.035957] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:15.570 [2024-10-29 11:05:21.035979] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:15.570 [2024-10-29 11:05:21.035995] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:15.570 [2024-10-29 11:05:21.038294] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:15.570 [2024-10-29 11:05:21.038342] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:15.571 [2024-10-29 11:05:21.038434] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:15.571 [2024-10-29 11:05:21.038507] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:15.571 [2024-10-29 11:05:21.038630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:15.571 spare 00:16:15.571 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.571 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:15.571 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.571 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:15.831 [2024-10-29 11:05:21.138551] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:16:15.831 [2024-10-29 11:05:21.138581] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:15.831 [2024-10-29 11:05:21.138847] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c19b0 00:16:15.831 [2024-10-29 11:05:21.138987] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:16:15.831 [2024-10-29 11:05:21.139002] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:16:15.831 [2024-10-29 11:05:21.139133] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.831 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:15.832 "name": "raid_bdev1", 00:16:15.832 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:15.832 "strip_size_kb": 0, 00:16:15.832 "state": "online", 00:16:15.832 "raid_level": "raid1", 00:16:15.832 "superblock": true, 00:16:15.832 "num_base_bdevs": 2, 00:16:15.832 "num_base_bdevs_discovered": 2, 00:16:15.832 "num_base_bdevs_operational": 2, 00:16:15.832 "base_bdevs_list": [ 00:16:15.832 { 00:16:15.832 "name": "spare", 00:16:15.832 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:15.832 "is_configured": true, 00:16:15.832 "data_offset": 256, 00:16:15.832 "data_size": 7936 00:16:15.832 }, 00:16:15.832 { 00:16:15.832 "name": "BaseBdev2", 00:16:15.832 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:15.832 "is_configured": true, 00:16:15.832 "data_offset": 256, 00:16:15.832 "data_size": 7936 00:16:15.832 } 00:16:15.832 ] 00:16:15.832 }' 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:15.832 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.403 "name": "raid_bdev1", 00:16:16.403 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:16.403 "strip_size_kb": 0, 00:16:16.403 "state": "online", 00:16:16.403 "raid_level": "raid1", 00:16:16.403 "superblock": true, 00:16:16.403 "num_base_bdevs": 2, 00:16:16.403 "num_base_bdevs_discovered": 2, 00:16:16.403 "num_base_bdevs_operational": 2, 00:16:16.403 "base_bdevs_list": [ 00:16:16.403 { 00:16:16.403 "name": "spare", 00:16:16.403 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:16.403 "is_configured": true, 00:16:16.403 "data_offset": 256, 00:16:16.403 "data_size": 7936 00:16:16.403 }, 00:16:16.403 { 00:16:16.403 "name": "BaseBdev2", 00:16:16.403 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:16.403 "is_configured": true, 00:16:16.403 "data_offset": 256, 00:16:16.403 "data_size": 7936 00:16:16.403 } 00:16:16.403 ] 00:16:16.403 }' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.403 [2024-10-29 11:05:21.798649] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.403 "name": "raid_bdev1", 00:16:16.403 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:16.403 "strip_size_kb": 0, 00:16:16.403 "state": "online", 00:16:16.403 "raid_level": "raid1", 00:16:16.403 "superblock": true, 00:16:16.403 "num_base_bdevs": 2, 00:16:16.403 "num_base_bdevs_discovered": 1, 00:16:16.403 "num_base_bdevs_operational": 1, 00:16:16.403 "base_bdevs_list": [ 00:16:16.403 { 00:16:16.403 "name": null, 00:16:16.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:16.403 "is_configured": false, 00:16:16.403 "data_offset": 0, 00:16:16.403 "data_size": 7936 00:16:16.403 }, 00:16:16.403 { 00:16:16.403 "name": "BaseBdev2", 00:16:16.403 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:16.403 "is_configured": true, 00:16:16.403 "data_offset": 256, 00:16:16.403 "data_size": 7936 00:16:16.403 } 00:16:16.403 ] 00:16:16.403 }' 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.403 11:05:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.974 11:05:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:16.974 11:05:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.974 11:05:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.974 [2024-10-29 11:05:22.281935] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:16.974 [2024-10-29 11:05:22.282142] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:16.974 [2024-10-29 11:05:22.282236] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:16.974 [2024-10-29 11:05:22.282326] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:16.974 [2024-10-29 11:05:22.287340] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1a80 00:16:16.974 11:05:22 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.974 11:05:22 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:16.974 [2024-10-29 11:05:22.289320] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:17.916 "name": "raid_bdev1", 00:16:17.916 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:17.916 "strip_size_kb": 0, 00:16:17.916 "state": "online", 00:16:17.916 "raid_level": "raid1", 00:16:17.916 "superblock": true, 00:16:17.916 "num_base_bdevs": 2, 00:16:17.916 "num_base_bdevs_discovered": 2, 00:16:17.916 "num_base_bdevs_operational": 2, 00:16:17.916 "process": { 00:16:17.916 "type": "rebuild", 00:16:17.916 "target": "spare", 00:16:17.916 "progress": { 00:16:17.916 "blocks": 2560, 00:16:17.916 "percent": 32 00:16:17.916 } 00:16:17.916 }, 00:16:17.916 "base_bdevs_list": [ 00:16:17.916 { 00:16:17.916 "name": "spare", 00:16:17.916 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:17.916 "is_configured": true, 00:16:17.916 "data_offset": 256, 00:16:17.916 "data_size": 7936 00:16:17.916 }, 00:16:17.916 { 00:16:17.916 "name": "BaseBdev2", 00:16:17.916 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:17.916 "is_configured": true, 00:16:17.916 "data_offset": 256, 00:16:17.916 "data_size": 7936 00:16:17.916 } 00:16:17.916 ] 00:16:17.916 }' 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:17.916 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.176 [2024-10-29 11:05:23.453563] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:18.176 [2024-10-29 11:05:23.493452] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:18.176 [2024-10-29 11:05:23.493569] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.176 [2024-10-29 11:05:23.493628] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:18.176 [2024-10-29 11:05:23.493668] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.176 "name": "raid_bdev1", 00:16:18.176 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:18.176 "strip_size_kb": 0, 00:16:18.176 "state": "online", 00:16:18.176 "raid_level": "raid1", 00:16:18.176 "superblock": true, 00:16:18.176 "num_base_bdevs": 2, 00:16:18.176 "num_base_bdevs_discovered": 1, 00:16:18.176 "num_base_bdevs_operational": 1, 00:16:18.176 "base_bdevs_list": [ 00:16:18.176 { 00:16:18.176 "name": null, 00:16:18.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.176 "is_configured": false, 00:16:18.176 "data_offset": 0, 00:16:18.176 "data_size": 7936 00:16:18.176 }, 00:16:18.176 { 00:16:18.176 "name": "BaseBdev2", 00:16:18.176 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:18.176 "is_configured": true, 00:16:18.176 "data_offset": 256, 00:16:18.176 "data_size": 7936 00:16:18.176 } 00:16:18.176 ] 00:16:18.176 }' 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.176 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.744 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:18.744 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.744 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.744 [2024-10-29 11:05:23.961097] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:18.744 [2024-10-29 11:05:23.961231] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.744 [2024-10-29 11:05:23.961276] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:18.744 [2024-10-29 11:05:23.961308] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.744 [2024-10-29 11:05:23.961825] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.744 [2024-10-29 11:05:23.961895] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:18.744 [2024-10-29 11:05:23.962008] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:18.744 [2024-10-29 11:05:23.962050] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:18.744 [2024-10-29 11:05:23.962123] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:18.744 [2024-10-29 11:05:23.962192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:18.744 [2024-10-29 11:05:23.966239] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:16:18.744 spare 00:16:18.744 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.744 11:05:23 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:18.744 [2024-10-29 11:05:23.968239] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.682 11:05:24 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.682 "name": "raid_bdev1", 00:16:19.682 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:19.682 "strip_size_kb": 0, 00:16:19.682 "state": "online", 00:16:19.682 "raid_level": "raid1", 00:16:19.682 "superblock": true, 00:16:19.682 "num_base_bdevs": 2, 00:16:19.682 "num_base_bdevs_discovered": 2, 00:16:19.682 "num_base_bdevs_operational": 2, 00:16:19.682 "process": { 00:16:19.682 "type": "rebuild", 00:16:19.682 "target": "spare", 00:16:19.682 "progress": { 00:16:19.682 "blocks": 2560, 00:16:19.682 "percent": 32 00:16:19.682 } 00:16:19.682 }, 00:16:19.682 "base_bdevs_list": [ 00:16:19.682 { 00:16:19.682 "name": "spare", 00:16:19.682 "uuid": "aa2aaca2-0a82-5cb0-b214-8e9b7d01b20e", 00:16:19.682 "is_configured": true, 00:16:19.682 "data_offset": 256, 00:16:19.682 "data_size": 7936 00:16:19.682 }, 00:16:19.682 { 00:16:19.682 "name": "BaseBdev2", 00:16:19.682 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:19.682 "is_configured": true, 00:16:19.682 "data_offset": 256, 00:16:19.682 "data_size": 7936 00:16:19.682 } 00:16:19.682 ] 00:16:19.682 }' 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.682 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.682 [2024-10-29 11:05:25.132530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:19.682 [2024-10-29 11:05:25.172332] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:19.682 [2024-10-29 11:05:25.172445] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:19.682 [2024-10-29 11:05:25.172463] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:19.682 [2024-10-29 11:05:25.172474] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.942 "name": "raid_bdev1", 00:16:19.942 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:19.942 "strip_size_kb": 0, 00:16:19.942 "state": "online", 00:16:19.942 "raid_level": "raid1", 00:16:19.942 "superblock": true, 00:16:19.942 "num_base_bdevs": 2, 00:16:19.942 "num_base_bdevs_discovered": 1, 00:16:19.942 "num_base_bdevs_operational": 1, 00:16:19.942 "base_bdevs_list": [ 00:16:19.942 { 00:16:19.942 "name": null, 00:16:19.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:19.942 "is_configured": false, 00:16:19.942 "data_offset": 0, 00:16:19.942 "data_size": 7936 00:16:19.942 }, 00:16:19.942 { 00:16:19.942 "name": "BaseBdev2", 00:16:19.942 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:19.942 "is_configured": true, 00:16:19.942 "data_offset": 256, 00:16:19.942 "data_size": 7936 00:16:19.942 } 00:16:19.942 ] 00:16:19.942 }' 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.942 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.201 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:20.461 "name": "raid_bdev1", 00:16:20.461 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:20.461 "strip_size_kb": 0, 00:16:20.461 "state": "online", 00:16:20.461 "raid_level": "raid1", 00:16:20.461 "superblock": true, 00:16:20.461 "num_base_bdevs": 2, 00:16:20.461 "num_base_bdevs_discovered": 1, 00:16:20.461 "num_base_bdevs_operational": 1, 00:16:20.461 "base_bdevs_list": [ 00:16:20.461 { 00:16:20.461 "name": null, 00:16:20.461 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.461 "is_configured": false, 00:16:20.461 "data_offset": 0, 00:16:20.461 "data_size": 7936 00:16:20.461 }, 00:16:20.461 { 00:16:20.461 "name": "BaseBdev2", 00:16:20.461 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:20.461 "is_configured": true, 00:16:20.461 "data_offset": 256, 00:16:20.461 "data_size": 7936 00:16:20.461 } 00:16:20.461 ] 00:16:20.461 }' 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.461 [2024-10-29 11:05:25.827743] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:20.461 [2024-10-29 11:05:25.827869] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.461 [2024-10-29 11:05:25.827894] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:20.461 [2024-10-29 11:05:25.827908] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.461 [2024-10-29 11:05:25.828337] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.461 [2024-10-29 11:05:25.828374] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:20.461 [2024-10-29 11:05:25.828450] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:20.461 [2024-10-29 11:05:25.828479] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:20.461 [2024-10-29 11:05:25.828500] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:20.461 [2024-10-29 11:05:25.828517] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:20.461 BaseBdev1 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.461 11:05:25 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:21.401 "name": "raid_bdev1", 00:16:21.401 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:21.401 "strip_size_kb": 0, 00:16:21.401 "state": "online", 00:16:21.401 "raid_level": "raid1", 00:16:21.401 "superblock": true, 00:16:21.401 "num_base_bdevs": 2, 00:16:21.401 "num_base_bdevs_discovered": 1, 00:16:21.401 "num_base_bdevs_operational": 1, 00:16:21.401 "base_bdevs_list": [ 00:16:21.401 { 00:16:21.401 "name": null, 00:16:21.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:21.401 "is_configured": false, 00:16:21.401 "data_offset": 0, 00:16:21.401 "data_size": 7936 00:16:21.401 }, 00:16:21.401 { 00:16:21.401 "name": "BaseBdev2", 00:16:21.401 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:21.401 "is_configured": true, 00:16:21.401 "data_offset": 256, 00:16:21.401 "data_size": 7936 00:16:21.401 } 00:16:21.401 ] 00:16:21.401 }' 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:21.401 11:05:26 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:21.972 "name": "raid_bdev1", 00:16:21.972 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:21.972 "strip_size_kb": 0, 00:16:21.972 "state": "online", 00:16:21.972 "raid_level": "raid1", 00:16:21.972 "superblock": true, 00:16:21.972 "num_base_bdevs": 2, 00:16:21.972 "num_base_bdevs_discovered": 1, 00:16:21.972 "num_base_bdevs_operational": 1, 00:16:21.972 "base_bdevs_list": [ 00:16:21.972 { 00:16:21.972 "name": null, 00:16:21.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:21.972 "is_configured": false, 00:16:21.972 "data_offset": 0, 00:16:21.972 "data_size": 7936 00:16:21.972 }, 00:16:21.972 { 00:16:21.972 "name": "BaseBdev2", 00:16:21.972 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:21.972 "is_configured": true, 00:16:21.972 "data_offset": 256, 00:16:21.972 "data_size": 7936 00:16:21.972 } 00:16:21.972 ] 00:16:21.972 }' 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # local es=0 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.972 [2024-10-29 11:05:27.409502] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:21.972 [2024-10-29 11:05:27.409656] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:21.972 [2024-10-29 11:05:27.409670] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:21.972 request: 00:16:21.972 { 00:16:21.972 "base_bdev": "BaseBdev1", 00:16:21.972 "raid_bdev": "raid_bdev1", 00:16:21.972 "method": "bdev_raid_add_base_bdev", 00:16:21.972 "req_id": 1 00:16:21.972 } 00:16:21.972 Got JSON-RPC error response 00:16:21.972 response: 00:16:21.972 { 00:16:21.972 "code": -22, 00:16:21.972 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:21.972 } 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # es=1 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:21.972 11:05:27 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.354 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.354 "name": "raid_bdev1", 00:16:23.354 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:23.354 "strip_size_kb": 0, 00:16:23.354 "state": "online", 00:16:23.354 "raid_level": "raid1", 00:16:23.354 "superblock": true, 00:16:23.354 "num_base_bdevs": 2, 00:16:23.354 "num_base_bdevs_discovered": 1, 00:16:23.354 "num_base_bdevs_operational": 1, 00:16:23.355 "base_bdevs_list": [ 00:16:23.355 { 00:16:23.355 "name": null, 00:16:23.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.355 "is_configured": false, 00:16:23.355 "data_offset": 0, 00:16:23.355 "data_size": 7936 00:16:23.355 }, 00:16:23.355 { 00:16:23.355 "name": "BaseBdev2", 00:16:23.355 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:23.355 "is_configured": true, 00:16:23.355 "data_offset": 256, 00:16:23.355 "data_size": 7936 00:16:23.355 } 00:16:23.355 ] 00:16:23.355 }' 00:16:23.355 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.355 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:23.614 "name": "raid_bdev1", 00:16:23.614 "uuid": "d8d069dc-9997-4660-9ddd-b5e13e1b9983", 00:16:23.614 "strip_size_kb": 0, 00:16:23.614 "state": "online", 00:16:23.614 "raid_level": "raid1", 00:16:23.614 "superblock": true, 00:16:23.614 "num_base_bdevs": 2, 00:16:23.614 "num_base_bdevs_discovered": 1, 00:16:23.614 "num_base_bdevs_operational": 1, 00:16:23.614 "base_bdevs_list": [ 00:16:23.614 { 00:16:23.614 "name": null, 00:16:23.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.614 "is_configured": false, 00:16:23.614 "data_offset": 0, 00:16:23.614 "data_size": 7936 00:16:23.614 }, 00:16:23.614 { 00:16:23.614 "name": "BaseBdev2", 00:16:23.614 "uuid": "ab04cfea-5040-5e68-983d-74e1d3f8b3c3", 00:16:23.614 "is_configured": true, 00:16:23.614 "data_offset": 256, 00:16:23.614 "data_size": 7936 00:16:23.614 } 00:16:23.614 ] 00:16:23.614 }' 00:16:23.614 11:05:28 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96980 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@952 -- # '[' -z 96980 ']' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # kill -0 96980 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@957 -- # uname 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 96980 00:16:23.614 killing process with pid 96980 00:16:23.614 Received shutdown signal, test time was about 60.000000 seconds 00:16:23.614 00:16:23.614 Latency(us) 00:16:23.614 [2024-10-29T11:05:29.112Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:23.614 [2024-10-29T11:05:29.112Z] =================================================================================================================== 00:16:23.614 [2024-10-29T11:05:29.112Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@970 -- # echo 'killing process with pid 96980' 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@971 -- # kill 96980 00:16:23.614 [2024-10-29 11:05:29.104247] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:23.614 [2024-10-29 11:05:29.104385] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:23.614 [2024-10-29 11:05:29.104442] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:23.614 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@976 -- # wait 96980 00:16:23.614 [2024-10-29 11:05:29.104452] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:16:23.873 [2024-10-29 11:05:29.135745] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:23.873 ************************************ 00:16:23.873 END TEST raid_rebuild_test_sb_4k 00:16:23.873 ************************************ 00:16:23.873 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:16:23.873 00:16:23.873 real 0m18.484s 00:16:23.873 user 0m24.441s 00:16:23.873 sys 0m2.780s 00:16:23.873 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:23.873 11:05:29 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:24.133 11:05:29 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:16:24.133 11:05:29 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:16:24.133 11:05:29 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:16:24.133 11:05:29 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:24.133 11:05:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:24.133 ************************************ 00:16:24.133 START TEST raid_state_function_test_sb_md_separate 00:16:24.133 ************************************ 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 2 true 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97655 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:24.133 Process raid pid: 97655 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97655' 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97655 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@833 -- # '[' -z 97655 ']' 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:24.133 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:24.133 11:05:29 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.133 [2024-10-29 11:05:29.511417] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:24.133 [2024-10-29 11:05:29.511540] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:24.393 [2024-10-29 11:05:29.680023] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:24.393 [2024-10-29 11:05:29.706141] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:24.393 [2024-10-29 11:05:29.750912] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:24.393 [2024-10-29 11:05:29.750956] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@866 -- # return 0 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.963 [2024-10-29 11:05:30.325321] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:24.963 [2024-10-29 11:05:30.325490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:24.963 [2024-10-29 11:05:30.325511] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:24.963 [2024-10-29 11:05:30.325523] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:24.963 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.964 "name": "Existed_Raid", 00:16:24.964 "uuid": "af15acf0-7902-4bbd-92ac-8e0c82c31500", 00:16:24.964 "strip_size_kb": 0, 00:16:24.964 "state": "configuring", 00:16:24.964 "raid_level": "raid1", 00:16:24.964 "superblock": true, 00:16:24.964 "num_base_bdevs": 2, 00:16:24.964 "num_base_bdevs_discovered": 0, 00:16:24.964 "num_base_bdevs_operational": 2, 00:16:24.964 "base_bdevs_list": [ 00:16:24.964 { 00:16:24.964 "name": "BaseBdev1", 00:16:24.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.964 "is_configured": false, 00:16:24.964 "data_offset": 0, 00:16:24.964 "data_size": 0 00:16:24.964 }, 00:16:24.964 { 00:16:24.964 "name": "BaseBdev2", 00:16:24.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.964 "is_configured": false, 00:16:24.964 "data_offset": 0, 00:16:24.964 "data_size": 0 00:16:24.964 } 00:16:24.964 ] 00:16:24.964 }' 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.964 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.535 [2024-10-29 11:05:30.832480] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:25.535 [2024-10-29 11:05:30.832593] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.535 [2024-10-29 11:05:30.844496] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:25.535 [2024-10-29 11:05:30.844609] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:25.535 [2024-10-29 11:05:30.844653] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:25.535 [2024-10-29 11:05:30.844681] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.535 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.536 [2024-10-29 11:05:30.866475] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:25.536 BaseBdev1 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local i 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.536 [ 00:16:25.536 { 00:16:25.536 "name": "BaseBdev1", 00:16:25.536 "aliases": [ 00:16:25.536 "7c7257f4-bb7a-4409-a73d-abdff2540d83" 00:16:25.536 ], 00:16:25.536 "product_name": "Malloc disk", 00:16:25.536 "block_size": 4096, 00:16:25.536 "num_blocks": 8192, 00:16:25.536 "uuid": "7c7257f4-bb7a-4409-a73d-abdff2540d83", 00:16:25.536 "md_size": 32, 00:16:25.536 "md_interleave": false, 00:16:25.536 "dif_type": 0, 00:16:25.536 "assigned_rate_limits": { 00:16:25.536 "rw_ios_per_sec": 0, 00:16:25.536 "rw_mbytes_per_sec": 0, 00:16:25.536 "r_mbytes_per_sec": 0, 00:16:25.536 "w_mbytes_per_sec": 0 00:16:25.536 }, 00:16:25.536 "claimed": true, 00:16:25.536 "claim_type": "exclusive_write", 00:16:25.536 "zoned": false, 00:16:25.536 "supported_io_types": { 00:16:25.536 "read": true, 00:16:25.536 "write": true, 00:16:25.536 "unmap": true, 00:16:25.536 "flush": true, 00:16:25.536 "reset": true, 00:16:25.536 "nvme_admin": false, 00:16:25.536 "nvme_io": false, 00:16:25.536 "nvme_io_md": false, 00:16:25.536 "write_zeroes": true, 00:16:25.536 "zcopy": true, 00:16:25.536 "get_zone_info": false, 00:16:25.536 "zone_management": false, 00:16:25.536 "zone_append": false, 00:16:25.536 "compare": false, 00:16:25.536 "compare_and_write": false, 00:16:25.536 "abort": true, 00:16:25.536 "seek_hole": false, 00:16:25.536 "seek_data": false, 00:16:25.536 "copy": true, 00:16:25.536 "nvme_iov_md": false 00:16:25.536 }, 00:16:25.536 "memory_domains": [ 00:16:25.536 { 00:16:25.536 "dma_device_id": "system", 00:16:25.536 "dma_device_type": 1 00:16:25.536 }, 00:16:25.536 { 00:16:25.536 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:25.536 "dma_device_type": 2 00:16:25.536 } 00:16:25.536 ], 00:16:25.536 "driver_specific": {} 00:16:25.536 } 00:16:25.536 ] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@909 -- # return 0 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:25.536 "name": "Existed_Raid", 00:16:25.536 "uuid": "2f7ee86e-c8f8-467e-a3db-bb85834ba05b", 00:16:25.536 "strip_size_kb": 0, 00:16:25.536 "state": "configuring", 00:16:25.536 "raid_level": "raid1", 00:16:25.536 "superblock": true, 00:16:25.536 "num_base_bdevs": 2, 00:16:25.536 "num_base_bdevs_discovered": 1, 00:16:25.536 "num_base_bdevs_operational": 2, 00:16:25.536 "base_bdevs_list": [ 00:16:25.536 { 00:16:25.536 "name": "BaseBdev1", 00:16:25.536 "uuid": "7c7257f4-bb7a-4409-a73d-abdff2540d83", 00:16:25.536 "is_configured": true, 00:16:25.536 "data_offset": 256, 00:16:25.536 "data_size": 7936 00:16:25.536 }, 00:16:25.536 { 00:16:25.536 "name": "BaseBdev2", 00:16:25.536 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:25.536 "is_configured": false, 00:16:25.536 "data_offset": 0, 00:16:25.536 "data_size": 0 00:16:25.536 } 00:16:25.536 ] 00:16:25.536 }' 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:25.536 11:05:30 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.107 [2024-10-29 11:05:31.337738] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:26.107 [2024-10-29 11:05:31.337862] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.107 [2024-10-29 11:05:31.349766] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:26.107 [2024-10-29 11:05:31.351695] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:26.107 [2024-10-29 11:05:31.351748] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:26.107 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.108 "name": "Existed_Raid", 00:16:26.108 "uuid": "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c", 00:16:26.108 "strip_size_kb": 0, 00:16:26.108 "state": "configuring", 00:16:26.108 "raid_level": "raid1", 00:16:26.108 "superblock": true, 00:16:26.108 "num_base_bdevs": 2, 00:16:26.108 "num_base_bdevs_discovered": 1, 00:16:26.108 "num_base_bdevs_operational": 2, 00:16:26.108 "base_bdevs_list": [ 00:16:26.108 { 00:16:26.108 "name": "BaseBdev1", 00:16:26.108 "uuid": "7c7257f4-bb7a-4409-a73d-abdff2540d83", 00:16:26.108 "is_configured": true, 00:16:26.108 "data_offset": 256, 00:16:26.108 "data_size": 7936 00:16:26.108 }, 00:16:26.108 { 00:16:26.108 "name": "BaseBdev2", 00:16:26.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.108 "is_configured": false, 00:16:26.108 "data_offset": 0, 00:16:26.108 "data_size": 0 00:16:26.108 } 00:16:26.108 ] 00:16:26.108 }' 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.108 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.368 [2024-10-29 11:05:31.828913] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:26.368 [2024-10-29 11:05:31.829270] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:26.368 [2024-10-29 11:05:31.829328] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:26.368 [2024-10-29 11:05:31.829512] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:26.368 BaseBdev2 00:16:26.368 [2024-10-29 11:05:31.829676] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:26.368 [2024-10-29 11:05:31.829694] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:16:26.368 [2024-10-29 11:05:31.829785] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local i 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.368 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.368 [ 00:16:26.368 { 00:16:26.368 "name": "BaseBdev2", 00:16:26.368 "aliases": [ 00:16:26.368 "fb8217ac-c455-45af-9cec-80ab70642dd1" 00:16:26.368 ], 00:16:26.368 "product_name": "Malloc disk", 00:16:26.368 "block_size": 4096, 00:16:26.368 "num_blocks": 8192, 00:16:26.368 "uuid": "fb8217ac-c455-45af-9cec-80ab70642dd1", 00:16:26.368 "md_size": 32, 00:16:26.368 "md_interleave": false, 00:16:26.368 "dif_type": 0, 00:16:26.368 "assigned_rate_limits": { 00:16:26.368 "rw_ios_per_sec": 0, 00:16:26.368 "rw_mbytes_per_sec": 0, 00:16:26.368 "r_mbytes_per_sec": 0, 00:16:26.368 "w_mbytes_per_sec": 0 00:16:26.368 }, 00:16:26.368 "claimed": true, 00:16:26.368 "claim_type": "exclusive_write", 00:16:26.369 "zoned": false, 00:16:26.369 "supported_io_types": { 00:16:26.369 "read": true, 00:16:26.369 "write": true, 00:16:26.369 "unmap": true, 00:16:26.369 "flush": true, 00:16:26.369 "reset": true, 00:16:26.369 "nvme_admin": false, 00:16:26.369 "nvme_io": false, 00:16:26.369 "nvme_io_md": false, 00:16:26.369 "write_zeroes": true, 00:16:26.369 "zcopy": true, 00:16:26.369 "get_zone_info": false, 00:16:26.369 "zone_management": false, 00:16:26.369 "zone_append": false, 00:16:26.369 "compare": false, 00:16:26.369 "compare_and_write": false, 00:16:26.369 "abort": true, 00:16:26.369 "seek_hole": false, 00:16:26.369 "seek_data": false, 00:16:26.369 "copy": true, 00:16:26.369 "nvme_iov_md": false 00:16:26.369 }, 00:16:26.369 "memory_domains": [ 00:16:26.369 { 00:16:26.369 "dma_device_id": "system", 00:16:26.369 "dma_device_type": 1 00:16:26.369 }, 00:16:26.369 { 00:16:26.369 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:26.369 "dma_device_type": 2 00:16:26.369 } 00:16:26.369 ], 00:16:26.369 "driver_specific": {} 00:16:26.369 } 00:16:26.369 ] 00:16:26.369 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.369 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@909 -- # return 0 00:16:26.369 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:26.369 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:26.369 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.628 "name": "Existed_Raid", 00:16:26.628 "uuid": "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c", 00:16:26.628 "strip_size_kb": 0, 00:16:26.628 "state": "online", 00:16:26.628 "raid_level": "raid1", 00:16:26.628 "superblock": true, 00:16:26.628 "num_base_bdevs": 2, 00:16:26.628 "num_base_bdevs_discovered": 2, 00:16:26.628 "num_base_bdevs_operational": 2, 00:16:26.628 "base_bdevs_list": [ 00:16:26.628 { 00:16:26.628 "name": "BaseBdev1", 00:16:26.628 "uuid": "7c7257f4-bb7a-4409-a73d-abdff2540d83", 00:16:26.628 "is_configured": true, 00:16:26.628 "data_offset": 256, 00:16:26.628 "data_size": 7936 00:16:26.628 }, 00:16:26.628 { 00:16:26.628 "name": "BaseBdev2", 00:16:26.628 "uuid": "fb8217ac-c455-45af-9cec-80ab70642dd1", 00:16:26.628 "is_configured": true, 00:16:26.628 "data_offset": 256, 00:16:26.628 "data_size": 7936 00:16:26.628 } 00:16:26.628 ] 00:16:26.628 }' 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.628 11:05:31 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.887 [2024-10-29 11:05:32.328562] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.887 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:26.887 "name": "Existed_Raid", 00:16:26.887 "aliases": [ 00:16:26.887 "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c" 00:16:26.887 ], 00:16:26.887 "product_name": "Raid Volume", 00:16:26.887 "block_size": 4096, 00:16:26.887 "num_blocks": 7936, 00:16:26.887 "uuid": "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c", 00:16:26.887 "md_size": 32, 00:16:26.887 "md_interleave": false, 00:16:26.887 "dif_type": 0, 00:16:26.887 "assigned_rate_limits": { 00:16:26.887 "rw_ios_per_sec": 0, 00:16:26.887 "rw_mbytes_per_sec": 0, 00:16:26.887 "r_mbytes_per_sec": 0, 00:16:26.887 "w_mbytes_per_sec": 0 00:16:26.887 }, 00:16:26.887 "claimed": false, 00:16:26.887 "zoned": false, 00:16:26.887 "supported_io_types": { 00:16:26.887 "read": true, 00:16:26.887 "write": true, 00:16:26.887 "unmap": false, 00:16:26.887 "flush": false, 00:16:26.887 "reset": true, 00:16:26.887 "nvme_admin": false, 00:16:26.887 "nvme_io": false, 00:16:26.887 "nvme_io_md": false, 00:16:26.887 "write_zeroes": true, 00:16:26.887 "zcopy": false, 00:16:26.887 "get_zone_info": false, 00:16:26.887 "zone_management": false, 00:16:26.887 "zone_append": false, 00:16:26.888 "compare": false, 00:16:26.888 "compare_and_write": false, 00:16:26.888 "abort": false, 00:16:26.888 "seek_hole": false, 00:16:26.888 "seek_data": false, 00:16:26.888 "copy": false, 00:16:26.888 "nvme_iov_md": false 00:16:26.888 }, 00:16:26.888 "memory_domains": [ 00:16:26.888 { 00:16:26.888 "dma_device_id": "system", 00:16:26.888 "dma_device_type": 1 00:16:26.888 }, 00:16:26.888 { 00:16:26.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:26.888 "dma_device_type": 2 00:16:26.888 }, 00:16:26.888 { 00:16:26.888 "dma_device_id": "system", 00:16:26.888 "dma_device_type": 1 00:16:26.888 }, 00:16:26.888 { 00:16:26.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:26.888 "dma_device_type": 2 00:16:26.888 } 00:16:26.888 ], 00:16:26.888 "driver_specific": { 00:16:26.888 "raid": { 00:16:26.888 "uuid": "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c", 00:16:26.888 "strip_size_kb": 0, 00:16:26.888 "state": "online", 00:16:26.888 "raid_level": "raid1", 00:16:26.888 "superblock": true, 00:16:26.888 "num_base_bdevs": 2, 00:16:26.888 "num_base_bdevs_discovered": 2, 00:16:26.888 "num_base_bdevs_operational": 2, 00:16:26.888 "base_bdevs_list": [ 00:16:26.888 { 00:16:26.888 "name": "BaseBdev1", 00:16:26.888 "uuid": "7c7257f4-bb7a-4409-a73d-abdff2540d83", 00:16:26.888 "is_configured": true, 00:16:26.888 "data_offset": 256, 00:16:26.888 "data_size": 7936 00:16:26.888 }, 00:16:26.888 { 00:16:26.888 "name": "BaseBdev2", 00:16:26.888 "uuid": "fb8217ac-c455-45af-9cec-80ab70642dd1", 00:16:26.888 "is_configured": true, 00:16:26.888 "data_offset": 256, 00:16:26.888 "data_size": 7936 00:16:26.888 } 00:16:26.888 ] 00:16:26.888 } 00:16:26.888 } 00:16:26.888 }' 00:16:26.888 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:27.148 BaseBdev2' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.148 [2024-10-29 11:05:32.572189] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:27.148 "name": "Existed_Raid", 00:16:27.148 "uuid": "68a30c4c-1b42-43b6-acd8-0f1876ee4c1c", 00:16:27.148 "strip_size_kb": 0, 00:16:27.148 "state": "online", 00:16:27.148 "raid_level": "raid1", 00:16:27.148 "superblock": true, 00:16:27.148 "num_base_bdevs": 2, 00:16:27.148 "num_base_bdevs_discovered": 1, 00:16:27.148 "num_base_bdevs_operational": 1, 00:16:27.148 "base_bdevs_list": [ 00:16:27.148 { 00:16:27.148 "name": null, 00:16:27.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.148 "is_configured": false, 00:16:27.148 "data_offset": 0, 00:16:27.148 "data_size": 7936 00:16:27.148 }, 00:16:27.148 { 00:16:27.148 "name": "BaseBdev2", 00:16:27.148 "uuid": "fb8217ac-c455-45af-9cec-80ab70642dd1", 00:16:27.148 "is_configured": true, 00:16:27.148 "data_offset": 256, 00:16:27.148 "data_size": 7936 00:16:27.148 } 00:16:27.148 ] 00:16:27.148 }' 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:27.148 11:05:32 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.719 [2024-10-29 11:05:33.076440] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:27.719 [2024-10-29 11:05:33.076626] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:27.719 [2024-10-29 11:05:33.089037] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:27.719 [2024-10-29 11:05:33.089174] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:27.719 [2024-10-29 11:05:33.089240] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97655 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@952 -- # '[' -z 97655 ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # kill -0 97655 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@957 -- # uname 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 97655 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:27.719 killing process with pid 97655 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@970 -- # echo 'killing process with pid 97655' 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@971 -- # kill 97655 00:16:27.719 [2024-10-29 11:05:33.185278] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:27.719 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@976 -- # wait 97655 00:16:27.719 [2024-10-29 11:05:33.186315] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:27.980 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:16:27.980 00:16:27.980 real 0m3.988s 00:16:27.980 user 0m6.242s 00:16:27.980 sys 0m0.897s 00:16:27.980 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:27.980 ************************************ 00:16:27.980 END TEST raid_state_function_test_sb_md_separate 00:16:27.980 ************************************ 00:16:27.980 11:05:33 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.980 11:05:33 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:16:27.980 11:05:33 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:16:27.980 11:05:33 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:27.980 11:05:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:28.241 ************************************ 00:16:28.241 START TEST raid_superblock_test_md_separate 00:16:28.241 ************************************ 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 2 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97896 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97896 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@833 -- # '[' -z 97896 ']' 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:28.241 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:28.241 11:05:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.241 [2024-10-29 11:05:33.577012] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:28.241 [2024-10-29 11:05:33.577228] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97896 ] 00:16:28.503 [2024-10-29 11:05:33.751978] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:28.503 [2024-10-29 11:05:33.778735] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:28.503 [2024-10-29 11:05:33.823171] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:28.503 [2024-10-29 11:05:33.823218] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@866 -- # return 0 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.075 malloc1 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.075 [2024-10-29 11:05:34.422836] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:29.075 [2024-10-29 11:05:34.423010] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:29.075 [2024-10-29 11:05:34.423053] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:29.075 [2024-10-29 11:05:34.423118] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:29.075 [2024-10-29 11:05:34.425190] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:29.075 [2024-10-29 11:05:34.425280] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:29.075 pt1 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:29.075 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.076 malloc2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.076 [2024-10-29 11:05:34.456243] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:29.076 [2024-10-29 11:05:34.456312] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:29.076 [2024-10-29 11:05:34.456329] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:29.076 [2024-10-29 11:05:34.456341] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:29.076 [2024-10-29 11:05:34.458273] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:29.076 [2024-10-29 11:05:34.458401] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:29.076 pt2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.076 [2024-10-29 11:05:34.468254] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:29.076 [2024-10-29 11:05:34.470145] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:29.076 [2024-10-29 11:05:34.470412] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:29.076 [2024-10-29 11:05:34.470436] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:29.076 [2024-10-29 11:05:34.470556] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:29.076 [2024-10-29 11:05:34.470657] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:29.076 [2024-10-29 11:05:34.470668] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:29.076 [2024-10-29 11:05:34.470775] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.076 "name": "raid_bdev1", 00:16:29.076 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:29.076 "strip_size_kb": 0, 00:16:29.076 "state": "online", 00:16:29.076 "raid_level": "raid1", 00:16:29.076 "superblock": true, 00:16:29.076 "num_base_bdevs": 2, 00:16:29.076 "num_base_bdevs_discovered": 2, 00:16:29.076 "num_base_bdevs_operational": 2, 00:16:29.076 "base_bdevs_list": [ 00:16:29.076 { 00:16:29.076 "name": "pt1", 00:16:29.076 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:29.076 "is_configured": true, 00:16:29.076 "data_offset": 256, 00:16:29.076 "data_size": 7936 00:16:29.076 }, 00:16:29.076 { 00:16:29.076 "name": "pt2", 00:16:29.076 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:29.076 "is_configured": true, 00:16:29.076 "data_offset": 256, 00:16:29.076 "data_size": 7936 00:16:29.076 } 00:16:29.076 ] 00:16:29.076 }' 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.076 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:29.648 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.649 11:05:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.649 [2024-10-29 11:05:34.992130] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:29.649 "name": "raid_bdev1", 00:16:29.649 "aliases": [ 00:16:29.649 "7cdc57bc-8871-440f-a952-f84701860b7f" 00:16:29.649 ], 00:16:29.649 "product_name": "Raid Volume", 00:16:29.649 "block_size": 4096, 00:16:29.649 "num_blocks": 7936, 00:16:29.649 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:29.649 "md_size": 32, 00:16:29.649 "md_interleave": false, 00:16:29.649 "dif_type": 0, 00:16:29.649 "assigned_rate_limits": { 00:16:29.649 "rw_ios_per_sec": 0, 00:16:29.649 "rw_mbytes_per_sec": 0, 00:16:29.649 "r_mbytes_per_sec": 0, 00:16:29.649 "w_mbytes_per_sec": 0 00:16:29.649 }, 00:16:29.649 "claimed": false, 00:16:29.649 "zoned": false, 00:16:29.649 "supported_io_types": { 00:16:29.649 "read": true, 00:16:29.649 "write": true, 00:16:29.649 "unmap": false, 00:16:29.649 "flush": false, 00:16:29.649 "reset": true, 00:16:29.649 "nvme_admin": false, 00:16:29.649 "nvme_io": false, 00:16:29.649 "nvme_io_md": false, 00:16:29.649 "write_zeroes": true, 00:16:29.649 "zcopy": false, 00:16:29.649 "get_zone_info": false, 00:16:29.649 "zone_management": false, 00:16:29.649 "zone_append": false, 00:16:29.649 "compare": false, 00:16:29.649 "compare_and_write": false, 00:16:29.649 "abort": false, 00:16:29.649 "seek_hole": false, 00:16:29.649 "seek_data": false, 00:16:29.649 "copy": false, 00:16:29.649 "nvme_iov_md": false 00:16:29.649 }, 00:16:29.649 "memory_domains": [ 00:16:29.649 { 00:16:29.649 "dma_device_id": "system", 00:16:29.649 "dma_device_type": 1 00:16:29.649 }, 00:16:29.649 { 00:16:29.649 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.649 "dma_device_type": 2 00:16:29.649 }, 00:16:29.649 { 00:16:29.649 "dma_device_id": "system", 00:16:29.649 "dma_device_type": 1 00:16:29.649 }, 00:16:29.649 { 00:16:29.649 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.649 "dma_device_type": 2 00:16:29.649 } 00:16:29.649 ], 00:16:29.649 "driver_specific": { 00:16:29.649 "raid": { 00:16:29.649 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:29.649 "strip_size_kb": 0, 00:16:29.649 "state": "online", 00:16:29.649 "raid_level": "raid1", 00:16:29.649 "superblock": true, 00:16:29.649 "num_base_bdevs": 2, 00:16:29.649 "num_base_bdevs_discovered": 2, 00:16:29.649 "num_base_bdevs_operational": 2, 00:16:29.649 "base_bdevs_list": [ 00:16:29.649 { 00:16:29.649 "name": "pt1", 00:16:29.649 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:29.649 "is_configured": true, 00:16:29.649 "data_offset": 256, 00:16:29.649 "data_size": 7936 00:16:29.649 }, 00:16:29.649 { 00:16:29.649 "name": "pt2", 00:16:29.649 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:29.649 "is_configured": true, 00:16:29.649 "data_offset": 256, 00:16:29.649 "data_size": 7936 00:16:29.649 } 00:16:29.649 ] 00:16:29.649 } 00:16:29.649 } 00:16:29.649 }' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:29.649 pt2' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.649 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.910 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 [2024-10-29 11:05:35.215693] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7cdc57bc-8871-440f-a952-f84701860b7f 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 7cdc57bc-8871-440f-a952-f84701860b7f ']' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 [2024-10-29 11:05:35.263393] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:29.911 [2024-10-29 11:05:35.263476] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:29.911 [2024-10-29 11:05:35.263579] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:29.911 [2024-10-29 11:05:35.263676] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:29.911 [2024-10-29 11:05:35.263763] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.911 [2024-10-29 11:05:35.399243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:29.911 [2024-10-29 11:05:35.401284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:29.911 [2024-10-29 11:05:35.401430] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:29.911 [2024-10-29 11:05:35.401522] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:29.911 [2024-10-29 11:05:35.401603] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:29.911 [2024-10-29 11:05:35.401640] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:16:29.911 request: 00:16:29.911 { 00:16:29.911 "name": "raid_bdev1", 00:16:29.911 "raid_level": "raid1", 00:16:29.911 "base_bdevs": [ 00:16:29.911 "malloc1", 00:16:29.911 "malloc2" 00:16:29.911 ], 00:16:29.911 "superblock": false, 00:16:29.911 "method": "bdev_raid_create", 00:16:29.911 "req_id": 1 00:16:29.911 } 00:16:29.911 Got JSON-RPC error response 00:16:29.911 response: 00:16:29.911 { 00:16:29.911 "code": -17, 00:16:29.911 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:29.911 } 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:29.911 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:30.172 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.173 [2024-10-29 11:05:35.463086] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:30.173 [2024-10-29 11:05:35.463202] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.173 [2024-10-29 11:05:35.463240] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:30.173 [2024-10-29 11:05:35.463271] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.173 [2024-10-29 11:05:35.465308] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.173 [2024-10-29 11:05:35.465397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:30.173 [2024-10-29 11:05:35.465470] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:30.173 [2024-10-29 11:05:35.465548] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:30.173 pt1 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:30.173 "name": "raid_bdev1", 00:16:30.173 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:30.173 "strip_size_kb": 0, 00:16:30.173 "state": "configuring", 00:16:30.173 "raid_level": "raid1", 00:16:30.173 "superblock": true, 00:16:30.173 "num_base_bdevs": 2, 00:16:30.173 "num_base_bdevs_discovered": 1, 00:16:30.173 "num_base_bdevs_operational": 2, 00:16:30.173 "base_bdevs_list": [ 00:16:30.173 { 00:16:30.173 "name": "pt1", 00:16:30.173 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:30.173 "is_configured": true, 00:16:30.173 "data_offset": 256, 00:16:30.173 "data_size": 7936 00:16:30.173 }, 00:16:30.173 { 00:16:30.173 "name": null, 00:16:30.173 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:30.173 "is_configured": false, 00:16:30.173 "data_offset": 256, 00:16:30.173 "data_size": 7936 00:16:30.173 } 00:16:30.173 ] 00:16:30.173 }' 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:30.173 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.434 [2024-10-29 11:05:35.918325] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:30.434 [2024-10-29 11:05:35.918472] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.434 [2024-10-29 11:05:35.918516] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:30.434 [2024-10-29 11:05:35.918548] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.434 [2024-10-29 11:05:35.918773] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.434 [2024-10-29 11:05:35.918831] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:30.434 [2024-10-29 11:05:35.918922] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:30.434 [2024-10-29 11:05:35.918969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:30.434 [2024-10-29 11:05:35.919095] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:30.434 [2024-10-29 11:05:35.919136] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:30.434 [2024-10-29 11:05:35.919255] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:30.434 [2024-10-29 11:05:35.919402] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:30.434 [2024-10-29 11:05:35.919423] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:16:30.434 [2024-10-29 11:05:35.919489] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:30.434 pt2 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.434 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:30.695 "name": "raid_bdev1", 00:16:30.695 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:30.695 "strip_size_kb": 0, 00:16:30.695 "state": "online", 00:16:30.695 "raid_level": "raid1", 00:16:30.695 "superblock": true, 00:16:30.695 "num_base_bdevs": 2, 00:16:30.695 "num_base_bdevs_discovered": 2, 00:16:30.695 "num_base_bdevs_operational": 2, 00:16:30.695 "base_bdevs_list": [ 00:16:30.695 { 00:16:30.695 "name": "pt1", 00:16:30.695 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:30.695 "is_configured": true, 00:16:30.695 "data_offset": 256, 00:16:30.695 "data_size": 7936 00:16:30.695 }, 00:16:30.695 { 00:16:30.695 "name": "pt2", 00:16:30.695 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:30.695 "is_configured": true, 00:16:30.695 "data_offset": 256, 00:16:30.695 "data_size": 7936 00:16:30.695 } 00:16:30.695 ] 00:16:30.695 }' 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:30.695 11:05:35 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.956 [2024-10-29 11:05:36.429712] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.956 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:30.956 "name": "raid_bdev1", 00:16:30.956 "aliases": [ 00:16:30.956 "7cdc57bc-8871-440f-a952-f84701860b7f" 00:16:30.956 ], 00:16:30.956 "product_name": "Raid Volume", 00:16:30.956 "block_size": 4096, 00:16:30.956 "num_blocks": 7936, 00:16:30.956 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:30.956 "md_size": 32, 00:16:30.956 "md_interleave": false, 00:16:30.956 "dif_type": 0, 00:16:30.956 "assigned_rate_limits": { 00:16:30.956 "rw_ios_per_sec": 0, 00:16:30.956 "rw_mbytes_per_sec": 0, 00:16:30.956 "r_mbytes_per_sec": 0, 00:16:30.956 "w_mbytes_per_sec": 0 00:16:30.956 }, 00:16:30.956 "claimed": false, 00:16:30.956 "zoned": false, 00:16:30.956 "supported_io_types": { 00:16:30.956 "read": true, 00:16:30.956 "write": true, 00:16:30.956 "unmap": false, 00:16:30.956 "flush": false, 00:16:30.956 "reset": true, 00:16:30.956 "nvme_admin": false, 00:16:30.956 "nvme_io": false, 00:16:30.956 "nvme_io_md": false, 00:16:30.956 "write_zeroes": true, 00:16:30.956 "zcopy": false, 00:16:30.956 "get_zone_info": false, 00:16:30.956 "zone_management": false, 00:16:30.956 "zone_append": false, 00:16:30.956 "compare": false, 00:16:30.956 "compare_and_write": false, 00:16:30.956 "abort": false, 00:16:30.956 "seek_hole": false, 00:16:30.956 "seek_data": false, 00:16:30.956 "copy": false, 00:16:30.956 "nvme_iov_md": false 00:16:30.956 }, 00:16:30.956 "memory_domains": [ 00:16:30.956 { 00:16:30.956 "dma_device_id": "system", 00:16:30.956 "dma_device_type": 1 00:16:30.956 }, 00:16:30.956 { 00:16:30.956 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.956 "dma_device_type": 2 00:16:30.956 }, 00:16:30.956 { 00:16:30.956 "dma_device_id": "system", 00:16:30.956 "dma_device_type": 1 00:16:30.956 }, 00:16:30.956 { 00:16:30.956 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.956 "dma_device_type": 2 00:16:30.956 } 00:16:30.956 ], 00:16:30.956 "driver_specific": { 00:16:30.956 "raid": { 00:16:30.956 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:30.956 "strip_size_kb": 0, 00:16:30.956 "state": "online", 00:16:30.956 "raid_level": "raid1", 00:16:30.956 "superblock": true, 00:16:30.956 "num_base_bdevs": 2, 00:16:30.956 "num_base_bdevs_discovered": 2, 00:16:30.956 "num_base_bdevs_operational": 2, 00:16:30.956 "base_bdevs_list": [ 00:16:30.956 { 00:16:30.956 "name": "pt1", 00:16:30.956 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:30.956 "is_configured": true, 00:16:30.956 "data_offset": 256, 00:16:30.956 "data_size": 7936 00:16:30.956 }, 00:16:30.957 { 00:16:30.957 "name": "pt2", 00:16:30.957 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:30.957 "is_configured": true, 00:16:30.957 "data_offset": 256, 00:16:30.957 "data_size": 7936 00:16:30.957 } 00:16:30.957 ] 00:16:30.957 } 00:16:30.957 } 00:16:30.957 }' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:31.217 pt2' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.217 [2024-10-29 11:05:36.645321] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.217 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 7cdc57bc-8871-440f-a952-f84701860b7f '!=' 7cdc57bc-8871-440f-a952-f84701860b7f ']' 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.218 [2024-10-29 11:05:36.689066] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.218 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.478 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.478 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.478 "name": "raid_bdev1", 00:16:31.478 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:31.478 "strip_size_kb": 0, 00:16:31.478 "state": "online", 00:16:31.478 "raid_level": "raid1", 00:16:31.478 "superblock": true, 00:16:31.478 "num_base_bdevs": 2, 00:16:31.478 "num_base_bdevs_discovered": 1, 00:16:31.478 "num_base_bdevs_operational": 1, 00:16:31.478 "base_bdevs_list": [ 00:16:31.478 { 00:16:31.478 "name": null, 00:16:31.478 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:31.478 "is_configured": false, 00:16:31.478 "data_offset": 0, 00:16:31.478 "data_size": 7936 00:16:31.478 }, 00:16:31.478 { 00:16:31.478 "name": "pt2", 00:16:31.478 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:31.478 "is_configured": true, 00:16:31.478 "data_offset": 256, 00:16:31.478 "data_size": 7936 00:16:31.478 } 00:16:31.478 ] 00:16:31.478 }' 00:16:31.478 11:05:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.478 11:05:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.738 [2024-10-29 11:05:37.168235] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:31.738 [2024-10-29 11:05:37.168322] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:31.738 [2024-10-29 11:05:37.168399] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:31.738 [2024-10-29 11:05:37.168475] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:31.738 [2024-10-29 11:05:37.168485] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.738 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.003 [2024-10-29 11:05:37.244157] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:32.003 [2024-10-29 11:05:37.244273] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:32.003 [2024-10-29 11:05:37.244297] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:32.003 [2024-10-29 11:05:37.244308] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:32.003 [2024-10-29 11:05:37.246334] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:32.003 [2024-10-29 11:05:37.246397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:32.003 [2024-10-29 11:05:37.246450] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:32.003 [2024-10-29 11:05:37.246482] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:32.003 [2024-10-29 11:05:37.246549] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:16:32.003 [2024-10-29 11:05:37.246558] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:32.003 [2024-10-29 11:05:37.246634] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:32.003 [2024-10-29 11:05:37.246720] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:16:32.003 [2024-10-29 11:05:37.246731] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:16:32.003 [2024-10-29 11:05:37.246822] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:32.003 pt2 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.003 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.004 "name": "raid_bdev1", 00:16:32.004 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:32.004 "strip_size_kb": 0, 00:16:32.004 "state": "online", 00:16:32.004 "raid_level": "raid1", 00:16:32.004 "superblock": true, 00:16:32.004 "num_base_bdevs": 2, 00:16:32.004 "num_base_bdevs_discovered": 1, 00:16:32.004 "num_base_bdevs_operational": 1, 00:16:32.004 "base_bdevs_list": [ 00:16:32.004 { 00:16:32.004 "name": null, 00:16:32.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.004 "is_configured": false, 00:16:32.004 "data_offset": 256, 00:16:32.004 "data_size": 7936 00:16:32.004 }, 00:16:32.004 { 00:16:32.004 "name": "pt2", 00:16:32.004 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.004 "is_configured": true, 00:16:32.004 "data_offset": 256, 00:16:32.004 "data_size": 7936 00:16:32.004 } 00:16:32.004 ] 00:16:32.004 }' 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.004 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 [2024-10-29 11:05:37.671853] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:32.266 [2024-10-29 11:05:37.671939] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:32.266 [2024-10-29 11:05:37.672053] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:32.266 [2024-10-29 11:05:37.672167] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:32.266 [2024-10-29 11:05:37.672229] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 [2024-10-29 11:05:37.731731] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:32.266 [2024-10-29 11:05:37.731859] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:32.266 [2024-10-29 11:05:37.731897] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:32.266 [2024-10-29 11:05:37.731964] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:32.266 [2024-10-29 11:05:37.734022] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:32.266 [2024-10-29 11:05:37.734105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:32.266 [2024-10-29 11:05:37.734179] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:32.266 [2024-10-29 11:05:37.734257] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:32.266 [2024-10-29 11:05:37.734418] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:32.266 [2024-10-29 11:05:37.734483] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:32.266 [2024-10-29 11:05:37.734549] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:16:32.266 [2024-10-29 11:05:37.734649] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:32.266 [2024-10-29 11:05:37.734749] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:16:32.266 [2024-10-29 11:05:37.734794] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:32.266 [2024-10-29 11:05:37.734915] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:32.266 [2024-10-29 11:05:37.735045] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:16:32.266 [2024-10-29 11:05:37.735087] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:16:32.266 [2024-10-29 11:05:37.735246] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:32.266 pt1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.526 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.526 "name": "raid_bdev1", 00:16:32.526 "uuid": "7cdc57bc-8871-440f-a952-f84701860b7f", 00:16:32.526 "strip_size_kb": 0, 00:16:32.526 "state": "online", 00:16:32.526 "raid_level": "raid1", 00:16:32.526 "superblock": true, 00:16:32.526 "num_base_bdevs": 2, 00:16:32.526 "num_base_bdevs_discovered": 1, 00:16:32.526 "num_base_bdevs_operational": 1, 00:16:32.526 "base_bdevs_list": [ 00:16:32.526 { 00:16:32.526 "name": null, 00:16:32.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.526 "is_configured": false, 00:16:32.526 "data_offset": 256, 00:16:32.526 "data_size": 7936 00:16:32.526 }, 00:16:32.526 { 00:16:32.526 "name": "pt2", 00:16:32.526 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.526 "is_configured": true, 00:16:32.526 "data_offset": 256, 00:16:32.526 "data_size": 7936 00:16:32.526 } 00:16:32.526 ] 00:16:32.526 }' 00:16:32.526 11:05:37 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.526 11:05:37 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.787 [2024-10-29 11:05:38.227412] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 7cdc57bc-8871-440f-a952-f84701860b7f '!=' 7cdc57bc-8871-440f-a952-f84701860b7f ']' 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97896 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@952 -- # '[' -z 97896 ']' 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # kill -0 97896 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@957 -- # uname 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:32.787 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 97896 00:16:33.047 killing process with pid 97896 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@970 -- # echo 'killing process with pid 97896' 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@971 -- # kill 97896 00:16:33.048 [2024-10-29 11:05:38.311629] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:33.048 [2024-10-29 11:05:38.311695] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:33.048 [2024-10-29 11:05:38.311735] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:33.048 [2024-10-29 11:05:38.311744] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@976 -- # wait 97896 00:16:33.048 [2024-10-29 11:05:38.336325] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:16:33.048 ************************************ 00:16:33.048 END TEST raid_superblock_test_md_separate 00:16:33.048 ************************************ 00:16:33.048 00:16:33.048 real 0m5.057s 00:16:33.048 user 0m8.251s 00:16:33.048 sys 0m1.136s 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:33.048 11:05:38 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.309 11:05:38 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:16:33.309 11:05:38 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:16:33.309 11:05:38 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:16:33.309 11:05:38 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:33.309 11:05:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:33.309 ************************************ 00:16:33.309 START TEST raid_rebuild_test_sb_md_separate 00:16:33.309 ************************************ 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 true false true 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=98210 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 98210 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@833 -- # '[' -z 98210 ']' 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:33.309 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:33.309 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:33.310 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:33.310 11:05:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.310 [2024-10-29 11:05:38.733002] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:33.310 [2024-10-29 11:05:38.733197] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98210 ] 00:16:33.310 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:33.310 Zero copy mechanism will not be used. 00:16:33.570 [2024-10-29 11:05:38.908035] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:33.570 [2024-10-29 11:05:38.934736] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:33.570 [2024-10-29 11:05:38.979054] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:33.570 [2024-10-29 11:05:38.979107] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@866 -- # return 0 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.141 BaseBdev1_malloc 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.141 [2024-10-29 11:05:39.586716] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:34.141 [2024-10-29 11:05:39.586785] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.141 [2024-10-29 11:05:39.586808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:34.141 [2024-10-29 11:05:39.586819] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.141 [2024-10-29 11:05:39.588766] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.141 [2024-10-29 11:05:39.588885] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:34.141 BaseBdev1 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.141 BaseBdev2_malloc 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.141 [2024-10-29 11:05:39.616158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:34.141 [2024-10-29 11:05:39.616310] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.141 [2024-10-29 11:05:39.616336] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:34.141 [2024-10-29 11:05:39.616347] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.141 [2024-10-29 11:05:39.618289] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.141 [2024-10-29 11:05:39.618330] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:34.141 BaseBdev2 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.141 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.400 spare_malloc 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.401 spare_delay 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.401 [2024-10-29 11:05:39.677953] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:34.401 [2024-10-29 11:05:39.678035] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.401 [2024-10-29 11:05:39.678076] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:34.401 [2024-10-29 11:05:39.678092] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.401 [2024-10-29 11:05:39.680854] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.401 [2024-10-29 11:05:39.680909] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:34.401 spare 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.401 [2024-10-29 11:05:39.689929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:34.401 [2024-10-29 11:05:39.691874] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:34.401 [2024-10-29 11:05:39.692167] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:34.401 [2024-10-29 11:05:39.692188] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:34.401 [2024-10-29 11:05:39.692289] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:34.401 [2024-10-29 11:05:39.692418] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:34.401 [2024-10-29 11:05:39.692432] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:34.401 [2024-10-29 11:05:39.692541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:34.401 "name": "raid_bdev1", 00:16:34.401 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:34.401 "strip_size_kb": 0, 00:16:34.401 "state": "online", 00:16:34.401 "raid_level": "raid1", 00:16:34.401 "superblock": true, 00:16:34.401 "num_base_bdevs": 2, 00:16:34.401 "num_base_bdevs_discovered": 2, 00:16:34.401 "num_base_bdevs_operational": 2, 00:16:34.401 "base_bdevs_list": [ 00:16:34.401 { 00:16:34.401 "name": "BaseBdev1", 00:16:34.401 "uuid": "13575df3-3008-5667-9f85-08220d945a87", 00:16:34.401 "is_configured": true, 00:16:34.401 "data_offset": 256, 00:16:34.401 "data_size": 7936 00:16:34.401 }, 00:16:34.401 { 00:16:34.401 "name": "BaseBdev2", 00:16:34.401 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:34.401 "is_configured": true, 00:16:34.401 "data_offset": 256, 00:16:34.401 "data_size": 7936 00:16:34.401 } 00:16:34.401 ] 00:16:34.401 }' 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:34.401 11:05:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.968 [2024-10-29 11:05:40.189284] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:34.968 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:34.969 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:34.969 [2024-10-29 11:05:40.448651] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:35.228 /dev/nbd0 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # local i 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # break 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:35.228 1+0 records in 00:16:35.228 1+0 records out 00:16:35.228 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000560065 s, 7.3 MB/s 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # size=4096 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # return 0 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:35.228 11:05:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:35.797 7936+0 records in 00:16:35.797 7936+0 records out 00:16:35.797 32505856 bytes (33 MB, 31 MiB) copied, 0.659893 s, 49.3 MB/s 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:35.797 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:36.057 [2024-10-29 11:05:41.400261] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.057 [2024-10-29 11:05:41.412562] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.057 "name": "raid_bdev1", 00:16:36.057 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:36.057 "strip_size_kb": 0, 00:16:36.057 "state": "online", 00:16:36.057 "raid_level": "raid1", 00:16:36.057 "superblock": true, 00:16:36.057 "num_base_bdevs": 2, 00:16:36.057 "num_base_bdevs_discovered": 1, 00:16:36.057 "num_base_bdevs_operational": 1, 00:16:36.057 "base_bdevs_list": [ 00:16:36.057 { 00:16:36.057 "name": null, 00:16:36.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.057 "is_configured": false, 00:16:36.057 "data_offset": 0, 00:16:36.057 "data_size": 7936 00:16:36.057 }, 00:16:36.057 { 00:16:36.057 "name": "BaseBdev2", 00:16:36.057 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:36.057 "is_configured": true, 00:16:36.057 "data_offset": 256, 00:16:36.057 "data_size": 7936 00:16:36.057 } 00:16:36.057 ] 00:16:36.057 }' 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.057 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.628 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:36.628 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.628 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.628 [2024-10-29 11:05:41.879962] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:36.628 [2024-10-29 11:05:41.882756] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d0c0 00:16:36.628 [2024-10-29 11:05:41.884688] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:36.628 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.628 11:05:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.568 "name": "raid_bdev1", 00:16:37.568 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:37.568 "strip_size_kb": 0, 00:16:37.568 "state": "online", 00:16:37.568 "raid_level": "raid1", 00:16:37.568 "superblock": true, 00:16:37.568 "num_base_bdevs": 2, 00:16:37.568 "num_base_bdevs_discovered": 2, 00:16:37.568 "num_base_bdevs_operational": 2, 00:16:37.568 "process": { 00:16:37.568 "type": "rebuild", 00:16:37.568 "target": "spare", 00:16:37.568 "progress": { 00:16:37.568 "blocks": 2560, 00:16:37.568 "percent": 32 00:16:37.568 } 00:16:37.568 }, 00:16:37.568 "base_bdevs_list": [ 00:16:37.568 { 00:16:37.568 "name": "spare", 00:16:37.568 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:37.568 "is_configured": true, 00:16:37.568 "data_offset": 256, 00:16:37.568 "data_size": 7936 00:16:37.568 }, 00:16:37.568 { 00:16:37.568 "name": "BaseBdev2", 00:16:37.568 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:37.568 "is_configured": true, 00:16:37.568 "data_offset": 256, 00:16:37.568 "data_size": 7936 00:16:37.568 } 00:16:37.568 ] 00:16:37.568 }' 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:37.568 11:05:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.568 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:37.568 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:37.568 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.568 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.568 [2024-10-29 11:05:43.040577] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:37.828 [2024-10-29 11:05:43.089526] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:37.828 [2024-10-29 11:05:43.089587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:37.828 [2024-10-29 11:05:43.089607] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:37.828 [2024-10-29 11:05:43.089615] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.828 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:37.828 "name": "raid_bdev1", 00:16:37.828 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:37.828 "strip_size_kb": 0, 00:16:37.828 "state": "online", 00:16:37.828 "raid_level": "raid1", 00:16:37.828 "superblock": true, 00:16:37.828 "num_base_bdevs": 2, 00:16:37.828 "num_base_bdevs_discovered": 1, 00:16:37.828 "num_base_bdevs_operational": 1, 00:16:37.828 "base_bdevs_list": [ 00:16:37.828 { 00:16:37.828 "name": null, 00:16:37.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.828 "is_configured": false, 00:16:37.828 "data_offset": 0, 00:16:37.829 "data_size": 7936 00:16:37.829 }, 00:16:37.829 { 00:16:37.829 "name": "BaseBdev2", 00:16:37.829 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:37.829 "is_configured": true, 00:16:37.829 "data_offset": 256, 00:16:37.829 "data_size": 7936 00:16:37.829 } 00:16:37.829 ] 00:16:37.829 }' 00:16:37.829 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:37.829 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:38.089 "name": "raid_bdev1", 00:16:38.089 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:38.089 "strip_size_kb": 0, 00:16:38.089 "state": "online", 00:16:38.089 "raid_level": "raid1", 00:16:38.089 "superblock": true, 00:16:38.089 "num_base_bdevs": 2, 00:16:38.089 "num_base_bdevs_discovered": 1, 00:16:38.089 "num_base_bdevs_operational": 1, 00:16:38.089 "base_bdevs_list": [ 00:16:38.089 { 00:16:38.089 "name": null, 00:16:38.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.089 "is_configured": false, 00:16:38.089 "data_offset": 0, 00:16:38.089 "data_size": 7936 00:16:38.089 }, 00:16:38.089 { 00:16:38.089 "name": "BaseBdev2", 00:16:38.089 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:38.089 "is_configured": true, 00:16:38.089 "data_offset": 256, 00:16:38.089 "data_size": 7936 00:16:38.089 } 00:16:38.089 ] 00:16:38.089 }' 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:38.089 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.349 [2024-10-29 11:05:43.624178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:38.349 [2024-10-29 11:05:43.626732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d190 00:16:38.349 [2024-10-29 11:05:43.628601] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.349 11:05:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.289 "name": "raid_bdev1", 00:16:39.289 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:39.289 "strip_size_kb": 0, 00:16:39.289 "state": "online", 00:16:39.289 "raid_level": "raid1", 00:16:39.289 "superblock": true, 00:16:39.289 "num_base_bdevs": 2, 00:16:39.289 "num_base_bdevs_discovered": 2, 00:16:39.289 "num_base_bdevs_operational": 2, 00:16:39.289 "process": { 00:16:39.289 "type": "rebuild", 00:16:39.289 "target": "spare", 00:16:39.289 "progress": { 00:16:39.289 "blocks": 2560, 00:16:39.289 "percent": 32 00:16:39.289 } 00:16:39.289 }, 00:16:39.289 "base_bdevs_list": [ 00:16:39.289 { 00:16:39.289 "name": "spare", 00:16:39.289 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:39.289 "is_configured": true, 00:16:39.289 "data_offset": 256, 00:16:39.289 "data_size": 7936 00:16:39.289 }, 00:16:39.289 { 00:16:39.289 "name": "BaseBdev2", 00:16:39.289 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:39.289 "is_configured": true, 00:16:39.289 "data_offset": 256, 00:16:39.289 "data_size": 7936 00:16:39.289 } 00:16:39.289 ] 00:16:39.289 }' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:39.289 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=589 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:39.289 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.556 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.556 "name": "raid_bdev1", 00:16:39.556 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:39.556 "strip_size_kb": 0, 00:16:39.556 "state": "online", 00:16:39.556 "raid_level": "raid1", 00:16:39.556 "superblock": true, 00:16:39.556 "num_base_bdevs": 2, 00:16:39.556 "num_base_bdevs_discovered": 2, 00:16:39.556 "num_base_bdevs_operational": 2, 00:16:39.556 "process": { 00:16:39.556 "type": "rebuild", 00:16:39.556 "target": "spare", 00:16:39.556 "progress": { 00:16:39.556 "blocks": 2816, 00:16:39.556 "percent": 35 00:16:39.556 } 00:16:39.556 }, 00:16:39.556 "base_bdevs_list": [ 00:16:39.556 { 00:16:39.556 "name": "spare", 00:16:39.556 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:39.556 "is_configured": true, 00:16:39.556 "data_offset": 256, 00:16:39.556 "data_size": 7936 00:16:39.556 }, 00:16:39.556 { 00:16:39.556 "name": "BaseBdev2", 00:16:39.557 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:39.557 "is_configured": true, 00:16:39.557 "data_offset": 256, 00:16:39.557 "data_size": 7936 00:16:39.557 } 00:16:39.557 ] 00:16:39.557 }' 00:16:39.557 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.557 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:39.557 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.557 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.557 11:05:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:40.554 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:40.555 "name": "raid_bdev1", 00:16:40.555 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:40.555 "strip_size_kb": 0, 00:16:40.555 "state": "online", 00:16:40.555 "raid_level": "raid1", 00:16:40.555 "superblock": true, 00:16:40.555 "num_base_bdevs": 2, 00:16:40.555 "num_base_bdevs_discovered": 2, 00:16:40.555 "num_base_bdevs_operational": 2, 00:16:40.555 "process": { 00:16:40.555 "type": "rebuild", 00:16:40.555 "target": "spare", 00:16:40.555 "progress": { 00:16:40.555 "blocks": 5632, 00:16:40.555 "percent": 70 00:16:40.555 } 00:16:40.555 }, 00:16:40.555 "base_bdevs_list": [ 00:16:40.555 { 00:16:40.555 "name": "spare", 00:16:40.555 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:40.555 "is_configured": true, 00:16:40.555 "data_offset": 256, 00:16:40.555 "data_size": 7936 00:16:40.555 }, 00:16:40.555 { 00:16:40.555 "name": "BaseBdev2", 00:16:40.555 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:40.555 "is_configured": true, 00:16:40.555 "data_offset": 256, 00:16:40.555 "data_size": 7936 00:16:40.555 } 00:16:40.555 ] 00:16:40.555 }' 00:16:40.555 11:05:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:40.555 11:05:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:40.555 11:05:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:40.815 11:05:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:40.815 11:05:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:41.386 [2024-10-29 11:05:46.739385] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:41.386 [2024-10-29 11:05:46.739535] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:41.386 [2024-10-29 11:05:46.739661] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.647 "name": "raid_bdev1", 00:16:41.647 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:41.647 "strip_size_kb": 0, 00:16:41.647 "state": "online", 00:16:41.647 "raid_level": "raid1", 00:16:41.647 "superblock": true, 00:16:41.647 "num_base_bdevs": 2, 00:16:41.647 "num_base_bdevs_discovered": 2, 00:16:41.647 "num_base_bdevs_operational": 2, 00:16:41.647 "base_bdevs_list": [ 00:16:41.647 { 00:16:41.647 "name": "spare", 00:16:41.647 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:41.647 "is_configured": true, 00:16:41.647 "data_offset": 256, 00:16:41.647 "data_size": 7936 00:16:41.647 }, 00:16:41.647 { 00:16:41.647 "name": "BaseBdev2", 00:16:41.647 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:41.647 "is_configured": true, 00:16:41.647 "data_offset": 256, 00:16:41.647 "data_size": 7936 00:16:41.647 } 00:16:41.647 ] 00:16:41.647 }' 00:16:41.647 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.907 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.908 "name": "raid_bdev1", 00:16:41.908 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:41.908 "strip_size_kb": 0, 00:16:41.908 "state": "online", 00:16:41.908 "raid_level": "raid1", 00:16:41.908 "superblock": true, 00:16:41.908 "num_base_bdevs": 2, 00:16:41.908 "num_base_bdevs_discovered": 2, 00:16:41.908 "num_base_bdevs_operational": 2, 00:16:41.908 "base_bdevs_list": [ 00:16:41.908 { 00:16:41.908 "name": "spare", 00:16:41.908 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:41.908 "is_configured": true, 00:16:41.908 "data_offset": 256, 00:16:41.908 "data_size": 7936 00:16:41.908 }, 00:16:41.908 { 00:16:41.908 "name": "BaseBdev2", 00:16:41.908 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:41.908 "is_configured": true, 00:16:41.908 "data_offset": 256, 00:16:41.908 "data_size": 7936 00:16:41.908 } 00:16:41.908 ] 00:16:41.908 }' 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:41.908 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.168 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:42.168 "name": "raid_bdev1", 00:16:42.168 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:42.168 "strip_size_kb": 0, 00:16:42.168 "state": "online", 00:16:42.168 "raid_level": "raid1", 00:16:42.168 "superblock": true, 00:16:42.168 "num_base_bdevs": 2, 00:16:42.168 "num_base_bdevs_discovered": 2, 00:16:42.168 "num_base_bdevs_operational": 2, 00:16:42.168 "base_bdevs_list": [ 00:16:42.168 { 00:16:42.168 "name": "spare", 00:16:42.168 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:42.168 "is_configured": true, 00:16:42.168 "data_offset": 256, 00:16:42.168 "data_size": 7936 00:16:42.168 }, 00:16:42.168 { 00:16:42.168 "name": "BaseBdev2", 00:16:42.168 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:42.168 "is_configured": true, 00:16:42.168 "data_offset": 256, 00:16:42.168 "data_size": 7936 00:16:42.168 } 00:16:42.168 ] 00:16:42.168 }' 00:16:42.168 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:42.168 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:42.428 [2024-10-29 11:05:47.849147] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:42.428 [2024-10-29 11:05:47.849247] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:42.428 [2024-10-29 11:05:47.849380] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:42.428 [2024-10-29 11:05:47.849452] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:42.428 [2024-10-29 11:05:47.849478] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:42.428 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:42.429 11:05:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:42.689 /dev/nbd0 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # local i 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # break 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:42.689 1+0 records in 00:16:42.689 1+0 records out 00:16:42.689 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000453457 s, 9.0 MB/s 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # size=4096 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # return 0 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:42.689 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:42.690 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:42.950 /dev/nbd1 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@870 -- # local nbd_name=nbd1 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # local i 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@874 -- # grep -q -w nbd1 /proc/partitions 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # break 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:42.950 1+0 records in 00:16:42.950 1+0 records out 00:16:42.950 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000353009 s, 11.6 MB/s 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # size=4096 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # return 0 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:42.950 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:43.211 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:43.471 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.472 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:43.732 [2024-10-29 11:05:48.973554] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:43.732 [2024-10-29 11:05:48.973712] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:43.732 [2024-10-29 11:05:48.973753] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:43.732 [2024-10-29 11:05:48.973811] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:43.732 [2024-10-29 11:05:48.975801] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:43.732 [2024-10-29 11:05:48.975907] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:43.732 [2024-10-29 11:05:48.975983] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:43.732 [2024-10-29 11:05:48.976045] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:43.732 [2024-10-29 11:05:48.976174] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:43.732 spare 00:16:43.732 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.732 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:43.732 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.732 11:05:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:43.732 [2024-10-29 11:05:49.076079] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:16:43.732 [2024-10-29 11:05:49.076107] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:43.732 [2024-10-29 11:05:49.076208] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c19b0 00:16:43.732 [2024-10-29 11:05:49.076326] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:16:43.732 [2024-10-29 11:05:49.076338] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:16:43.732 [2024-10-29 11:05:49.076476] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.732 "name": "raid_bdev1", 00:16:43.732 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:43.732 "strip_size_kb": 0, 00:16:43.732 "state": "online", 00:16:43.732 "raid_level": "raid1", 00:16:43.732 "superblock": true, 00:16:43.732 "num_base_bdevs": 2, 00:16:43.732 "num_base_bdevs_discovered": 2, 00:16:43.732 "num_base_bdevs_operational": 2, 00:16:43.732 "base_bdevs_list": [ 00:16:43.732 { 00:16:43.732 "name": "spare", 00:16:43.732 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:43.732 "is_configured": true, 00:16:43.732 "data_offset": 256, 00:16:43.732 "data_size": 7936 00:16:43.732 }, 00:16:43.732 { 00:16:43.732 "name": "BaseBdev2", 00:16:43.732 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:43.732 "is_configured": true, 00:16:43.732 "data_offset": 256, 00:16:43.732 "data_size": 7936 00:16:43.732 } 00:16:43.732 ] 00:16:43.732 }' 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.732 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:44.307 "name": "raid_bdev1", 00:16:44.307 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:44.307 "strip_size_kb": 0, 00:16:44.307 "state": "online", 00:16:44.307 "raid_level": "raid1", 00:16:44.307 "superblock": true, 00:16:44.307 "num_base_bdevs": 2, 00:16:44.307 "num_base_bdevs_discovered": 2, 00:16:44.307 "num_base_bdevs_operational": 2, 00:16:44.307 "base_bdevs_list": [ 00:16:44.307 { 00:16:44.307 "name": "spare", 00:16:44.307 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:44.307 "is_configured": true, 00:16:44.307 "data_offset": 256, 00:16:44.307 "data_size": 7936 00:16:44.307 }, 00:16:44.307 { 00:16:44.307 "name": "BaseBdev2", 00:16:44.307 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:44.307 "is_configured": true, 00:16:44.307 "data_offset": 256, 00:16:44.307 "data_size": 7936 00:16:44.307 } 00:16:44.307 ] 00:16:44.307 }' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.307 [2024-10-29 11:05:49.708452] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:44.307 "name": "raid_bdev1", 00:16:44.307 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:44.307 "strip_size_kb": 0, 00:16:44.307 "state": "online", 00:16:44.307 "raid_level": "raid1", 00:16:44.307 "superblock": true, 00:16:44.307 "num_base_bdevs": 2, 00:16:44.307 "num_base_bdevs_discovered": 1, 00:16:44.307 "num_base_bdevs_operational": 1, 00:16:44.307 "base_bdevs_list": [ 00:16:44.307 { 00:16:44.307 "name": null, 00:16:44.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:44.307 "is_configured": false, 00:16:44.307 "data_offset": 0, 00:16:44.307 "data_size": 7936 00:16:44.307 }, 00:16:44.307 { 00:16:44.307 "name": "BaseBdev2", 00:16:44.307 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:44.307 "is_configured": true, 00:16:44.307 "data_offset": 256, 00:16:44.307 "data_size": 7936 00:16:44.307 } 00:16:44.307 ] 00:16:44.307 }' 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:44.307 11:05:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.877 11:05:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:44.877 11:05:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.877 11:05:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.877 [2024-10-29 11:05:50.136176] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:44.877 [2024-10-29 11:05:50.136438] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:44.877 [2024-10-29 11:05:50.136506] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:44.877 [2024-10-29 11:05:50.136596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:44.877 [2024-10-29 11:05:50.139147] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1a80 00:16:44.877 [2024-10-29 11:05:50.141057] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:44.877 11:05:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.877 11:05:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:45.819 "name": "raid_bdev1", 00:16:45.819 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:45.819 "strip_size_kb": 0, 00:16:45.819 "state": "online", 00:16:45.819 "raid_level": "raid1", 00:16:45.819 "superblock": true, 00:16:45.819 "num_base_bdevs": 2, 00:16:45.819 "num_base_bdevs_discovered": 2, 00:16:45.819 "num_base_bdevs_operational": 2, 00:16:45.819 "process": { 00:16:45.819 "type": "rebuild", 00:16:45.819 "target": "spare", 00:16:45.819 "progress": { 00:16:45.819 "blocks": 2560, 00:16:45.819 "percent": 32 00:16:45.819 } 00:16:45.819 }, 00:16:45.819 "base_bdevs_list": [ 00:16:45.819 { 00:16:45.819 "name": "spare", 00:16:45.819 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:45.819 "is_configured": true, 00:16:45.819 "data_offset": 256, 00:16:45.819 "data_size": 7936 00:16:45.819 }, 00:16:45.819 { 00:16:45.819 "name": "BaseBdev2", 00:16:45.819 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:45.819 "is_configured": true, 00:16:45.819 "data_offset": 256, 00:16:45.819 "data_size": 7936 00:16:45.819 } 00:16:45.819 ] 00:16:45.819 }' 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.819 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.819 [2024-10-29 11:05:51.288250] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:46.080 [2024-10-29 11:05:51.345346] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:46.080 [2024-10-29 11:05:51.345413] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:46.080 [2024-10-29 11:05:51.345432] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:46.080 [2024-10-29 11:05:51.345440] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:46.080 "name": "raid_bdev1", 00:16:46.080 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:46.080 "strip_size_kb": 0, 00:16:46.080 "state": "online", 00:16:46.080 "raid_level": "raid1", 00:16:46.080 "superblock": true, 00:16:46.080 "num_base_bdevs": 2, 00:16:46.080 "num_base_bdevs_discovered": 1, 00:16:46.080 "num_base_bdevs_operational": 1, 00:16:46.080 "base_bdevs_list": [ 00:16:46.080 { 00:16:46.080 "name": null, 00:16:46.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.080 "is_configured": false, 00:16:46.080 "data_offset": 0, 00:16:46.080 "data_size": 7936 00:16:46.080 }, 00:16:46.080 { 00:16:46.080 "name": "BaseBdev2", 00:16:46.080 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:46.080 "is_configured": true, 00:16:46.080 "data_offset": 256, 00:16:46.080 "data_size": 7936 00:16:46.080 } 00:16:46.080 ] 00:16:46.080 }' 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:46.080 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.340 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:46.340 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.340 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.340 [2024-10-29 11:05:51.795668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:46.340 [2024-10-29 11:05:51.795731] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:46.340 [2024-10-29 11:05:51.795755] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:16:46.340 [2024-10-29 11:05:51.795766] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:46.340 [2024-10-29 11:05:51.795983] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:46.340 [2024-10-29 11:05:51.795999] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:46.340 [2024-10-29 11:05:51.796069] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:46.340 [2024-10-29 11:05:51.796081] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:46.340 [2024-10-29 11:05:51.796106] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:46.340 [2024-10-29 11:05:51.796129] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:46.340 [2024-10-29 11:05:51.798333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:16:46.340 [2024-10-29 11:05:51.800172] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:46.340 spare 00:16:46.340 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.340 11:05:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:47.722 "name": "raid_bdev1", 00:16:47.722 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:47.722 "strip_size_kb": 0, 00:16:47.722 "state": "online", 00:16:47.722 "raid_level": "raid1", 00:16:47.722 "superblock": true, 00:16:47.722 "num_base_bdevs": 2, 00:16:47.722 "num_base_bdevs_discovered": 2, 00:16:47.722 "num_base_bdevs_operational": 2, 00:16:47.722 "process": { 00:16:47.722 "type": "rebuild", 00:16:47.722 "target": "spare", 00:16:47.722 "progress": { 00:16:47.722 "blocks": 2560, 00:16:47.722 "percent": 32 00:16:47.722 } 00:16:47.722 }, 00:16:47.722 "base_bdevs_list": [ 00:16:47.722 { 00:16:47.722 "name": "spare", 00:16:47.722 "uuid": "cee3e61b-70e1-5cfc-9d04-4a996be62278", 00:16:47.722 "is_configured": true, 00:16:47.722 "data_offset": 256, 00:16:47.722 "data_size": 7936 00:16:47.722 }, 00:16:47.722 { 00:16:47.722 "name": "BaseBdev2", 00:16:47.722 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:47.722 "is_configured": true, 00:16:47.722 "data_offset": 256, 00:16:47.722 "data_size": 7936 00:16:47.722 } 00:16:47.722 ] 00:16:47.722 }' 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:47.722 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:47.723 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:47.723 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.723 11:05:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.723 [2024-10-29 11:05:52.963071] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.723 [2024-10-29 11:05:53.004426] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:47.723 [2024-10-29 11:05:53.004490] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:47.723 [2024-10-29 11:05:53.004506] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.723 [2024-10-29 11:05:53.004516] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:47.723 "name": "raid_bdev1", 00:16:47.723 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:47.723 "strip_size_kb": 0, 00:16:47.723 "state": "online", 00:16:47.723 "raid_level": "raid1", 00:16:47.723 "superblock": true, 00:16:47.723 "num_base_bdevs": 2, 00:16:47.723 "num_base_bdevs_discovered": 1, 00:16:47.723 "num_base_bdevs_operational": 1, 00:16:47.723 "base_bdevs_list": [ 00:16:47.723 { 00:16:47.723 "name": null, 00:16:47.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.723 "is_configured": false, 00:16:47.723 "data_offset": 0, 00:16:47.723 "data_size": 7936 00:16:47.723 }, 00:16:47.723 { 00:16:47.723 "name": "BaseBdev2", 00:16:47.723 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:47.723 "is_configured": true, 00:16:47.723 "data_offset": 256, 00:16:47.723 "data_size": 7936 00:16:47.723 } 00:16:47.723 ] 00:16:47.723 }' 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:47.723 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.983 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.243 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:48.244 "name": "raid_bdev1", 00:16:48.244 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:48.244 "strip_size_kb": 0, 00:16:48.244 "state": "online", 00:16:48.244 "raid_level": "raid1", 00:16:48.244 "superblock": true, 00:16:48.244 "num_base_bdevs": 2, 00:16:48.244 "num_base_bdevs_discovered": 1, 00:16:48.244 "num_base_bdevs_operational": 1, 00:16:48.244 "base_bdevs_list": [ 00:16:48.244 { 00:16:48.244 "name": null, 00:16:48.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.244 "is_configured": false, 00:16:48.244 "data_offset": 0, 00:16:48.244 "data_size": 7936 00:16:48.244 }, 00:16:48.244 { 00:16:48.244 "name": "BaseBdev2", 00:16:48.244 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:48.244 "is_configured": true, 00:16:48.244 "data_offset": 256, 00:16:48.244 "data_size": 7936 00:16:48.244 } 00:16:48.244 ] 00:16:48.244 }' 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:48.244 [2024-10-29 11:05:53.590401] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:48.244 [2024-10-29 11:05:53.590525] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:48.244 [2024-10-29 11:05:53.590549] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:16:48.244 [2024-10-29 11:05:53.590562] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:48.244 [2024-10-29 11:05:53.590771] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:48.244 [2024-10-29 11:05:53.590788] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:48.244 [2024-10-29 11:05:53.590842] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:48.244 [2024-10-29 11:05:53.590864] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:48.244 [2024-10-29 11:05:53.590873] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:48.244 [2024-10-29 11:05:53.590887] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:48.244 BaseBdev1 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.244 11:05:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.185 "name": "raid_bdev1", 00:16:49.185 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:49.185 "strip_size_kb": 0, 00:16:49.185 "state": "online", 00:16:49.185 "raid_level": "raid1", 00:16:49.185 "superblock": true, 00:16:49.185 "num_base_bdevs": 2, 00:16:49.185 "num_base_bdevs_discovered": 1, 00:16:49.185 "num_base_bdevs_operational": 1, 00:16:49.185 "base_bdevs_list": [ 00:16:49.185 { 00:16:49.185 "name": null, 00:16:49.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.185 "is_configured": false, 00:16:49.185 "data_offset": 0, 00:16:49.185 "data_size": 7936 00:16:49.185 }, 00:16:49.185 { 00:16:49.185 "name": "BaseBdev2", 00:16:49.185 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:49.185 "is_configured": true, 00:16:49.185 "data_offset": 256, 00:16:49.185 "data_size": 7936 00:16:49.185 } 00:16:49.185 ] 00:16:49.185 }' 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.185 11:05:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:49.755 "name": "raid_bdev1", 00:16:49.755 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:49.755 "strip_size_kb": 0, 00:16:49.755 "state": "online", 00:16:49.755 "raid_level": "raid1", 00:16:49.755 "superblock": true, 00:16:49.755 "num_base_bdevs": 2, 00:16:49.755 "num_base_bdevs_discovered": 1, 00:16:49.755 "num_base_bdevs_operational": 1, 00:16:49.755 "base_bdevs_list": [ 00:16:49.755 { 00:16:49.755 "name": null, 00:16:49.755 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.755 "is_configured": false, 00:16:49.755 "data_offset": 0, 00:16:49.755 "data_size": 7936 00:16:49.755 }, 00:16:49.755 { 00:16:49.755 "name": "BaseBdev2", 00:16:49.755 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:49.755 "is_configured": true, 00:16:49.755 "data_offset": 256, 00:16:49.755 "data_size": 7936 00:16:49.755 } 00:16:49.755 ] 00:16:49.755 }' 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:49.755 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.756 [2024-10-29 11:05:55.211738] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:49.756 [2024-10-29 11:05:55.211981] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:49.756 [2024-10-29 11:05:55.212053] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:49.756 request: 00:16:49.756 { 00:16:49.756 "base_bdev": "BaseBdev1", 00:16:49.756 "raid_bdev": "raid_bdev1", 00:16:49.756 "method": "bdev_raid_add_base_bdev", 00:16:49.756 "req_id": 1 00:16:49.756 } 00:16:49.756 Got JSON-RPC error response 00:16:49.756 response: 00:16:49.756 { 00:16:49.756 "code": -22, 00:16:49.756 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:49.756 } 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:49.756 11:05:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:51.137 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:51.137 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:51.138 "name": "raid_bdev1", 00:16:51.138 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:51.138 "strip_size_kb": 0, 00:16:51.138 "state": "online", 00:16:51.138 "raid_level": "raid1", 00:16:51.138 "superblock": true, 00:16:51.138 "num_base_bdevs": 2, 00:16:51.138 "num_base_bdevs_discovered": 1, 00:16:51.138 "num_base_bdevs_operational": 1, 00:16:51.138 "base_bdevs_list": [ 00:16:51.138 { 00:16:51.138 "name": null, 00:16:51.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:51.138 "is_configured": false, 00:16:51.138 "data_offset": 0, 00:16:51.138 "data_size": 7936 00:16:51.138 }, 00:16:51.138 { 00:16:51.138 "name": "BaseBdev2", 00:16:51.138 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:51.138 "is_configured": true, 00:16:51.138 "data_offset": 256, 00:16:51.138 "data_size": 7936 00:16:51.138 } 00:16:51.138 ] 00:16:51.138 }' 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:51.138 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:51.398 "name": "raid_bdev1", 00:16:51.398 "uuid": "d474e330-22bb-4a39-8e02-1ddb31b89d36", 00:16:51.398 "strip_size_kb": 0, 00:16:51.398 "state": "online", 00:16:51.398 "raid_level": "raid1", 00:16:51.398 "superblock": true, 00:16:51.398 "num_base_bdevs": 2, 00:16:51.398 "num_base_bdevs_discovered": 1, 00:16:51.398 "num_base_bdevs_operational": 1, 00:16:51.398 "base_bdevs_list": [ 00:16:51.398 { 00:16:51.398 "name": null, 00:16:51.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:51.398 "is_configured": false, 00:16:51.398 "data_offset": 0, 00:16:51.398 "data_size": 7936 00:16:51.398 }, 00:16:51.398 { 00:16:51.398 "name": "BaseBdev2", 00:16:51.398 "uuid": "767f5ec6-5120-59df-a65b-120e115d0de8", 00:16:51.398 "is_configured": true, 00:16:51.398 "data_offset": 256, 00:16:51.398 "data_size": 7936 00:16:51.398 } 00:16:51.398 ] 00:16:51.398 }' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 98210 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@952 -- # '[' -z 98210 ']' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # kill -0 98210 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@957 -- # uname 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:51.398 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 98210 00:16:51.658 killing process with pid 98210 00:16:51.658 Received shutdown signal, test time was about 60.000000 seconds 00:16:51.658 00:16:51.658 Latency(us) 00:16:51.658 [2024-10-29T11:05:57.156Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:51.658 [2024-10-29T11:05:57.156Z] =================================================================================================================== 00:16:51.658 [2024-10-29T11:05:57.156Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:51.658 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:51.658 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:51.658 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@970 -- # echo 'killing process with pid 98210' 00:16:51.658 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@971 -- # kill 98210 00:16:51.658 [2024-10-29 11:05:56.902330] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:51.658 [2024-10-29 11:05:56.902473] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:51.658 11:05:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@976 -- # wait 98210 00:16:51.658 [2024-10-29 11:05:56.902529] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:51.658 [2024-10-29 11:05:56.902541] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:16:51.658 [2024-10-29 11:05:56.935592] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:51.659 11:05:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:51.659 00:16:51.659 real 0m18.512s 00:16:51.659 user 0m24.507s 00:16:51.659 sys 0m2.740s 00:16:51.659 11:05:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:51.659 ************************************ 00:16:51.659 END TEST raid_rebuild_test_sb_md_separate 00:16:51.659 ************************************ 00:16:51.659 11:05:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.919 11:05:57 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:51.919 11:05:57 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:51.919 11:05:57 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:16:51.919 11:05:57 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:51.919 11:05:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:51.919 ************************************ 00:16:51.919 START TEST raid_state_function_test_sb_md_interleaved 00:16:51.919 ************************************ 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1127 -- # raid_state_function_test raid1 2 true 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98890 00:16:51.919 Process raid pid: 98890 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98890' 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98890 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@833 -- # '[' -z 98890 ']' 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:51.919 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:51.919 11:05:57 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:51.919 [2024-10-29 11:05:57.313690] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:51.919 [2024-10-29 11:05:57.313901] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:52.179 [2024-10-29 11:05:57.483168] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:52.179 [2024-10-29 11:05:57.509429] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:52.179 [2024-10-29 11:05:57.553420] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:52.179 [2024-10-29 11:05:57.553470] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@866 -- # return 0 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:52.750 [2024-10-29 11:05:58.131713] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:52.750 [2024-10-29 11:05:58.131776] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:52.750 [2024-10-29 11:05:58.131796] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:52.750 [2024-10-29 11:05:58.131810] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:52.750 "name": "Existed_Raid", 00:16:52.750 "uuid": "a6d5dcd1-632f-4b14-8b99-419187614562", 00:16:52.750 "strip_size_kb": 0, 00:16:52.750 "state": "configuring", 00:16:52.750 "raid_level": "raid1", 00:16:52.750 "superblock": true, 00:16:52.750 "num_base_bdevs": 2, 00:16:52.750 "num_base_bdevs_discovered": 0, 00:16:52.750 "num_base_bdevs_operational": 2, 00:16:52.750 "base_bdevs_list": [ 00:16:52.750 { 00:16:52.750 "name": "BaseBdev1", 00:16:52.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:52.750 "is_configured": false, 00:16:52.750 "data_offset": 0, 00:16:52.750 "data_size": 0 00:16:52.750 }, 00:16:52.750 { 00:16:52.750 "name": "BaseBdev2", 00:16:52.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:52.750 "is_configured": false, 00:16:52.750 "data_offset": 0, 00:16:52.750 "data_size": 0 00:16:52.750 } 00:16:52.750 ] 00:16:52.750 }' 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:52.750 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 [2024-10-29 11:05:58.638740] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:53.321 [2024-10-29 11:05:58.638786] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name Existed_Raid, state configuring 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 [2024-10-29 11:05:58.650728] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:53.321 [2024-10-29 11:05:58.650777] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:53.321 [2024-10-29 11:05:58.650786] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:53.321 [2024-10-29 11:05:58.650797] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 [2024-10-29 11:05:58.671943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:53.321 BaseBdev1 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev1 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local i 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 [ 00:16:53.321 { 00:16:53.321 "name": "BaseBdev1", 00:16:53.321 "aliases": [ 00:16:53.321 "cee6b952-8725-4a31-b4a6-63c79b249e1a" 00:16:53.321 ], 00:16:53.321 "product_name": "Malloc disk", 00:16:53.321 "block_size": 4128, 00:16:53.321 "num_blocks": 8192, 00:16:53.321 "uuid": "cee6b952-8725-4a31-b4a6-63c79b249e1a", 00:16:53.321 "md_size": 32, 00:16:53.321 "md_interleave": true, 00:16:53.321 "dif_type": 0, 00:16:53.321 "assigned_rate_limits": { 00:16:53.321 "rw_ios_per_sec": 0, 00:16:53.321 "rw_mbytes_per_sec": 0, 00:16:53.321 "r_mbytes_per_sec": 0, 00:16:53.321 "w_mbytes_per_sec": 0 00:16:53.321 }, 00:16:53.321 "claimed": true, 00:16:53.321 "claim_type": "exclusive_write", 00:16:53.321 "zoned": false, 00:16:53.321 "supported_io_types": { 00:16:53.321 "read": true, 00:16:53.321 "write": true, 00:16:53.321 "unmap": true, 00:16:53.321 "flush": true, 00:16:53.321 "reset": true, 00:16:53.321 "nvme_admin": false, 00:16:53.321 "nvme_io": false, 00:16:53.321 "nvme_io_md": false, 00:16:53.321 "write_zeroes": true, 00:16:53.321 "zcopy": true, 00:16:53.321 "get_zone_info": false, 00:16:53.321 "zone_management": false, 00:16:53.321 "zone_append": false, 00:16:53.321 "compare": false, 00:16:53.321 "compare_and_write": false, 00:16:53.321 "abort": true, 00:16:53.321 "seek_hole": false, 00:16:53.321 "seek_data": false, 00:16:53.321 "copy": true, 00:16:53.321 "nvme_iov_md": false 00:16:53.321 }, 00:16:53.321 "memory_domains": [ 00:16:53.321 { 00:16:53.321 "dma_device_id": "system", 00:16:53.321 "dma_device_type": 1 00:16:53.321 }, 00:16:53.321 { 00:16:53.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:53.321 "dma_device_type": 2 00:16:53.321 } 00:16:53.321 ], 00:16:53.321 "driver_specific": {} 00:16:53.321 } 00:16:53.321 ] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@909 -- # return 0 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.321 "name": "Existed_Raid", 00:16:53.321 "uuid": "aa46d91b-ebb2-4680-8f95-9663549bed71", 00:16:53.321 "strip_size_kb": 0, 00:16:53.321 "state": "configuring", 00:16:53.321 "raid_level": "raid1", 00:16:53.321 "superblock": true, 00:16:53.321 "num_base_bdevs": 2, 00:16:53.321 "num_base_bdevs_discovered": 1, 00:16:53.321 "num_base_bdevs_operational": 2, 00:16:53.321 "base_bdevs_list": [ 00:16:53.321 { 00:16:53.321 "name": "BaseBdev1", 00:16:53.321 "uuid": "cee6b952-8725-4a31-b4a6-63c79b249e1a", 00:16:53.321 "is_configured": true, 00:16:53.321 "data_offset": 256, 00:16:53.321 "data_size": 7936 00:16:53.321 }, 00:16:53.321 { 00:16:53.321 "name": "BaseBdev2", 00:16:53.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:53.321 "is_configured": false, 00:16:53.321 "data_offset": 0, 00:16:53.321 "data_size": 0 00:16:53.321 } 00:16:53.321 ] 00:16:53.321 }' 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.321 11:05:58 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.891 [2024-10-29 11:05:59.123324] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:53.891 [2024-10-29 11:05:59.123390] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name Existed_Raid, state configuring 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.891 [2024-10-29 11:05:59.131328] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:53.891 [2024-10-29 11:05:59.133191] bdev.c:8271:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:53.891 [2024-10-29 11:05:59.133242] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.891 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:53.891 "name": "Existed_Raid", 00:16:53.891 "uuid": "df65028f-056f-466b-a586-7a09f8e1d93e", 00:16:53.891 "strip_size_kb": 0, 00:16:53.891 "state": "configuring", 00:16:53.891 "raid_level": "raid1", 00:16:53.891 "superblock": true, 00:16:53.892 "num_base_bdevs": 2, 00:16:53.892 "num_base_bdevs_discovered": 1, 00:16:53.892 "num_base_bdevs_operational": 2, 00:16:53.892 "base_bdevs_list": [ 00:16:53.892 { 00:16:53.892 "name": "BaseBdev1", 00:16:53.892 "uuid": "cee6b952-8725-4a31-b4a6-63c79b249e1a", 00:16:53.892 "is_configured": true, 00:16:53.892 "data_offset": 256, 00:16:53.892 "data_size": 7936 00:16:53.892 }, 00:16:53.892 { 00:16:53.892 "name": "BaseBdev2", 00:16:53.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:53.892 "is_configured": false, 00:16:53.892 "data_offset": 0, 00:16:53.892 "data_size": 0 00:16:53.892 } 00:16:53.892 ] 00:16:53.892 }' 00:16:53.892 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:53.892 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.152 [2024-10-29 11:05:59.581954] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:54.152 [2024-10-29 11:05:59.582294] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:54.152 [2024-10-29 11:05:59.582352] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:54.152 [2024-10-29 11:05:59.582551] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:54.152 BaseBdev2 00:16:54.152 [2024-10-29 11:05:59.582671] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:54.152 [2024-10-29 11:05:59.582690] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000006980 00:16:54.152 [2024-10-29 11:05:59.582766] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local bdev_name=BaseBdev2 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # local bdev_timeout= 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local i 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # [[ -z '' ]] 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # bdev_timeout=2000 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_wait_for_examine 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.152 [ 00:16:54.152 { 00:16:54.152 "name": "BaseBdev2", 00:16:54.152 "aliases": [ 00:16:54.152 "57fe19b1-23f6-411d-81ad-8806dc1ba81e" 00:16:54.152 ], 00:16:54.152 "product_name": "Malloc disk", 00:16:54.152 "block_size": 4128, 00:16:54.152 "num_blocks": 8192, 00:16:54.152 "uuid": "57fe19b1-23f6-411d-81ad-8806dc1ba81e", 00:16:54.152 "md_size": 32, 00:16:54.152 "md_interleave": true, 00:16:54.152 "dif_type": 0, 00:16:54.152 "assigned_rate_limits": { 00:16:54.152 "rw_ios_per_sec": 0, 00:16:54.152 "rw_mbytes_per_sec": 0, 00:16:54.152 "r_mbytes_per_sec": 0, 00:16:54.152 "w_mbytes_per_sec": 0 00:16:54.152 }, 00:16:54.152 "claimed": true, 00:16:54.152 "claim_type": "exclusive_write", 00:16:54.152 "zoned": false, 00:16:54.152 "supported_io_types": { 00:16:54.152 "read": true, 00:16:54.152 "write": true, 00:16:54.152 "unmap": true, 00:16:54.152 "flush": true, 00:16:54.152 "reset": true, 00:16:54.152 "nvme_admin": false, 00:16:54.152 "nvme_io": false, 00:16:54.152 "nvme_io_md": false, 00:16:54.152 "write_zeroes": true, 00:16:54.152 "zcopy": true, 00:16:54.152 "get_zone_info": false, 00:16:54.152 "zone_management": false, 00:16:54.152 "zone_append": false, 00:16:54.152 "compare": false, 00:16:54.152 "compare_and_write": false, 00:16:54.152 "abort": true, 00:16:54.152 "seek_hole": false, 00:16:54.152 "seek_data": false, 00:16:54.152 "copy": true, 00:16:54.152 "nvme_iov_md": false 00:16:54.152 }, 00:16:54.152 "memory_domains": [ 00:16:54.152 { 00:16:54.152 "dma_device_id": "system", 00:16:54.152 "dma_device_type": 1 00:16:54.152 }, 00:16:54.152 { 00:16:54.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:54.152 "dma_device_type": 2 00:16:54.152 } 00:16:54.152 ], 00:16:54.152 "driver_specific": {} 00:16:54.152 } 00:16:54.152 ] 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@909 -- # return 0 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:54.152 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.153 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.153 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.412 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:54.412 "name": "Existed_Raid", 00:16:54.412 "uuid": "df65028f-056f-466b-a586-7a09f8e1d93e", 00:16:54.412 "strip_size_kb": 0, 00:16:54.412 "state": "online", 00:16:54.412 "raid_level": "raid1", 00:16:54.412 "superblock": true, 00:16:54.412 "num_base_bdevs": 2, 00:16:54.412 "num_base_bdevs_discovered": 2, 00:16:54.412 "num_base_bdevs_operational": 2, 00:16:54.412 "base_bdevs_list": [ 00:16:54.412 { 00:16:54.412 "name": "BaseBdev1", 00:16:54.412 "uuid": "cee6b952-8725-4a31-b4a6-63c79b249e1a", 00:16:54.412 "is_configured": true, 00:16:54.412 "data_offset": 256, 00:16:54.412 "data_size": 7936 00:16:54.412 }, 00:16:54.412 { 00:16:54.412 "name": "BaseBdev2", 00:16:54.412 "uuid": "57fe19b1-23f6-411d-81ad-8806dc1ba81e", 00:16:54.412 "is_configured": true, 00:16:54.412 "data_offset": 256, 00:16:54.412 "data_size": 7936 00:16:54.412 } 00:16:54.412 ] 00:16:54.412 }' 00:16:54.412 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:54.412 11:05:59 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.672 [2024-10-29 11:06:00.085425] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:54.672 "name": "Existed_Raid", 00:16:54.672 "aliases": [ 00:16:54.672 "df65028f-056f-466b-a586-7a09f8e1d93e" 00:16:54.672 ], 00:16:54.672 "product_name": "Raid Volume", 00:16:54.672 "block_size": 4128, 00:16:54.672 "num_blocks": 7936, 00:16:54.672 "uuid": "df65028f-056f-466b-a586-7a09f8e1d93e", 00:16:54.672 "md_size": 32, 00:16:54.672 "md_interleave": true, 00:16:54.672 "dif_type": 0, 00:16:54.672 "assigned_rate_limits": { 00:16:54.672 "rw_ios_per_sec": 0, 00:16:54.672 "rw_mbytes_per_sec": 0, 00:16:54.672 "r_mbytes_per_sec": 0, 00:16:54.672 "w_mbytes_per_sec": 0 00:16:54.672 }, 00:16:54.672 "claimed": false, 00:16:54.672 "zoned": false, 00:16:54.672 "supported_io_types": { 00:16:54.672 "read": true, 00:16:54.672 "write": true, 00:16:54.672 "unmap": false, 00:16:54.672 "flush": false, 00:16:54.672 "reset": true, 00:16:54.672 "nvme_admin": false, 00:16:54.672 "nvme_io": false, 00:16:54.672 "nvme_io_md": false, 00:16:54.672 "write_zeroes": true, 00:16:54.672 "zcopy": false, 00:16:54.672 "get_zone_info": false, 00:16:54.672 "zone_management": false, 00:16:54.672 "zone_append": false, 00:16:54.672 "compare": false, 00:16:54.672 "compare_and_write": false, 00:16:54.672 "abort": false, 00:16:54.672 "seek_hole": false, 00:16:54.672 "seek_data": false, 00:16:54.672 "copy": false, 00:16:54.672 "nvme_iov_md": false 00:16:54.672 }, 00:16:54.672 "memory_domains": [ 00:16:54.672 { 00:16:54.672 "dma_device_id": "system", 00:16:54.672 "dma_device_type": 1 00:16:54.672 }, 00:16:54.672 { 00:16:54.672 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:54.672 "dma_device_type": 2 00:16:54.672 }, 00:16:54.672 { 00:16:54.672 "dma_device_id": "system", 00:16:54.672 "dma_device_type": 1 00:16:54.672 }, 00:16:54.672 { 00:16:54.672 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:54.672 "dma_device_type": 2 00:16:54.672 } 00:16:54.672 ], 00:16:54.672 "driver_specific": { 00:16:54.672 "raid": { 00:16:54.672 "uuid": "df65028f-056f-466b-a586-7a09f8e1d93e", 00:16:54.672 "strip_size_kb": 0, 00:16:54.672 "state": "online", 00:16:54.672 "raid_level": "raid1", 00:16:54.672 "superblock": true, 00:16:54.672 "num_base_bdevs": 2, 00:16:54.672 "num_base_bdevs_discovered": 2, 00:16:54.672 "num_base_bdevs_operational": 2, 00:16:54.672 "base_bdevs_list": [ 00:16:54.672 { 00:16:54.672 "name": "BaseBdev1", 00:16:54.672 "uuid": "cee6b952-8725-4a31-b4a6-63c79b249e1a", 00:16:54.672 "is_configured": true, 00:16:54.672 "data_offset": 256, 00:16:54.672 "data_size": 7936 00:16:54.672 }, 00:16:54.672 { 00:16:54.672 "name": "BaseBdev2", 00:16:54.672 "uuid": "57fe19b1-23f6-411d-81ad-8806dc1ba81e", 00:16:54.672 "is_configured": true, 00:16:54.672 "data_offset": 256, 00:16:54.672 "data_size": 7936 00:16:54.672 } 00:16:54.672 ] 00:16:54.672 } 00:16:54.672 } 00:16:54.672 }' 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:54.672 BaseBdev2' 00:16:54.672 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.933 [2024-10-29 11:06:00.292836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:54.933 "name": "Existed_Raid", 00:16:54.933 "uuid": "df65028f-056f-466b-a586-7a09f8e1d93e", 00:16:54.933 "strip_size_kb": 0, 00:16:54.933 "state": "online", 00:16:54.933 "raid_level": "raid1", 00:16:54.933 "superblock": true, 00:16:54.933 "num_base_bdevs": 2, 00:16:54.933 "num_base_bdevs_discovered": 1, 00:16:54.933 "num_base_bdevs_operational": 1, 00:16:54.933 "base_bdevs_list": [ 00:16:54.933 { 00:16:54.933 "name": null, 00:16:54.933 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.933 "is_configured": false, 00:16:54.933 "data_offset": 0, 00:16:54.933 "data_size": 7936 00:16:54.933 }, 00:16:54.933 { 00:16:54.933 "name": "BaseBdev2", 00:16:54.933 "uuid": "57fe19b1-23f6-411d-81ad-8806dc1ba81e", 00:16:54.933 "is_configured": true, 00:16:54.933 "data_offset": 256, 00:16:54.933 "data_size": 7936 00:16:54.933 } 00:16:54.933 ] 00:16:54.933 }' 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:54.933 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.504 [2024-10-29 11:06:00.832216] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:55.504 [2024-10-29 11:06:00.832324] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:55.504 [2024-10-29 11:06:00.844224] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:55.504 [2024-10-29 11:06:00.844347] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:55.504 [2024-10-29 11:06:00.844413] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name Existed_Raid, state offline 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98890 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@952 -- # '[' -z 98890 ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # kill -0 98890 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@957 -- # uname 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 98890 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:16:55.504 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@970 -- # echo 'killing process with pid 98890' 00:16:55.504 killing process with pid 98890 00:16:55.505 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@971 -- # kill 98890 00:16:55.505 [2024-10-29 11:06:00.942867] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:55.505 11:06:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@976 -- # wait 98890 00:16:55.505 [2024-10-29 11:06:00.943878] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:55.765 11:06:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:55.765 00:16:55.765 real 0m3.931s 00:16:55.765 user 0m6.191s 00:16:55.765 sys 0m0.867s 00:16:55.765 11:06:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1128 -- # xtrace_disable 00:16:55.765 ************************************ 00:16:55.765 END TEST raid_state_function_test_sb_md_interleaved 00:16:55.765 ************************************ 00:16:55.765 11:06:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.765 11:06:01 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:55.765 11:06:01 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 4 -le 1 ']' 00:16:55.765 11:06:01 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:16:55.765 11:06:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:55.765 ************************************ 00:16:55.765 START TEST raid_superblock_test_md_interleaved 00:16:55.765 ************************************ 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1127 -- # raid_superblock_test raid1 2 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:55.765 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=99131 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 99131 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@833 -- # '[' -z 99131 ']' 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # local max_retries=100 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:55.766 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # xtrace_disable 00:16:55.766 11:06:01 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.026 [2024-10-29 11:06:01.321638] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:16:56.026 [2024-10-29 11:06:01.321775] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99131 ] 00:16:56.026 [2024-10-29 11:06:01.497396] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:56.026 [2024-10-29 11:06:01.524006] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:56.287 [2024-10-29 11:06:01.568243] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:56.287 [2024-10-29 11:06:01.568285] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@866 -- # return 0 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.858 malloc1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.858 [2024-10-29 11:06:02.143427] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:56.858 [2024-10-29 11:06:02.143580] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:56.858 [2024-10-29 11:06:02.143631] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:56.858 [2024-10-29 11:06:02.143696] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:56.858 [2024-10-29 11:06:02.145710] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:56.858 [2024-10-29 11:06:02.145794] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:56.858 pt1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.858 malloc2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.858 [2024-10-29 11:06:02.176358] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:56.858 [2024-10-29 11:06:02.176436] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:56.858 [2024-10-29 11:06:02.176456] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:56.858 [2024-10-29 11:06:02.176468] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:56.858 [2024-10-29 11:06:02.178331] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:56.858 [2024-10-29 11:06:02.178392] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:56.858 pt2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.858 [2024-10-29 11:06:02.188382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:56.858 [2024-10-29 11:06:02.190233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:56.858 [2024-10-29 11:06:02.190425] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:16:56.858 [2024-10-29 11:06:02.190444] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:56.858 [2024-10-29 11:06:02.190528] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:16:56.858 [2024-10-29 11:06:02.190600] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:16:56.858 [2024-10-29 11:06:02.190618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:16:56.858 [2024-10-29 11:06:02.190712] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:56.858 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:56.859 "name": "raid_bdev1", 00:16:56.859 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:56.859 "strip_size_kb": 0, 00:16:56.859 "state": "online", 00:16:56.859 "raid_level": "raid1", 00:16:56.859 "superblock": true, 00:16:56.859 "num_base_bdevs": 2, 00:16:56.859 "num_base_bdevs_discovered": 2, 00:16:56.859 "num_base_bdevs_operational": 2, 00:16:56.859 "base_bdevs_list": [ 00:16:56.859 { 00:16:56.859 "name": "pt1", 00:16:56.859 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:56.859 "is_configured": true, 00:16:56.859 "data_offset": 256, 00:16:56.859 "data_size": 7936 00:16:56.859 }, 00:16:56.859 { 00:16:56.859 "name": "pt2", 00:16:56.859 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:56.859 "is_configured": true, 00:16:56.859 "data_offset": 256, 00:16:56.859 "data_size": 7936 00:16:56.859 } 00:16:56.859 ] 00:16:56.859 }' 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:56.859 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.119 [2024-10-29 11:06:02.596045] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:57.119 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:57.380 "name": "raid_bdev1", 00:16:57.380 "aliases": [ 00:16:57.380 "58571275-c117-4c19-a0b8-962295c41dda" 00:16:57.380 ], 00:16:57.380 "product_name": "Raid Volume", 00:16:57.380 "block_size": 4128, 00:16:57.380 "num_blocks": 7936, 00:16:57.380 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:57.380 "md_size": 32, 00:16:57.380 "md_interleave": true, 00:16:57.380 "dif_type": 0, 00:16:57.380 "assigned_rate_limits": { 00:16:57.380 "rw_ios_per_sec": 0, 00:16:57.380 "rw_mbytes_per_sec": 0, 00:16:57.380 "r_mbytes_per_sec": 0, 00:16:57.380 "w_mbytes_per_sec": 0 00:16:57.380 }, 00:16:57.380 "claimed": false, 00:16:57.380 "zoned": false, 00:16:57.380 "supported_io_types": { 00:16:57.380 "read": true, 00:16:57.380 "write": true, 00:16:57.380 "unmap": false, 00:16:57.380 "flush": false, 00:16:57.380 "reset": true, 00:16:57.380 "nvme_admin": false, 00:16:57.380 "nvme_io": false, 00:16:57.380 "nvme_io_md": false, 00:16:57.380 "write_zeroes": true, 00:16:57.380 "zcopy": false, 00:16:57.380 "get_zone_info": false, 00:16:57.380 "zone_management": false, 00:16:57.380 "zone_append": false, 00:16:57.380 "compare": false, 00:16:57.380 "compare_and_write": false, 00:16:57.380 "abort": false, 00:16:57.380 "seek_hole": false, 00:16:57.380 "seek_data": false, 00:16:57.380 "copy": false, 00:16:57.380 "nvme_iov_md": false 00:16:57.380 }, 00:16:57.380 "memory_domains": [ 00:16:57.380 { 00:16:57.380 "dma_device_id": "system", 00:16:57.380 "dma_device_type": 1 00:16:57.380 }, 00:16:57.380 { 00:16:57.380 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:57.380 "dma_device_type": 2 00:16:57.380 }, 00:16:57.380 { 00:16:57.380 "dma_device_id": "system", 00:16:57.380 "dma_device_type": 1 00:16:57.380 }, 00:16:57.380 { 00:16:57.380 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:57.380 "dma_device_type": 2 00:16:57.380 } 00:16:57.380 ], 00:16:57.380 "driver_specific": { 00:16:57.380 "raid": { 00:16:57.380 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:57.380 "strip_size_kb": 0, 00:16:57.380 "state": "online", 00:16:57.380 "raid_level": "raid1", 00:16:57.380 "superblock": true, 00:16:57.380 "num_base_bdevs": 2, 00:16:57.380 "num_base_bdevs_discovered": 2, 00:16:57.380 "num_base_bdevs_operational": 2, 00:16:57.380 "base_bdevs_list": [ 00:16:57.380 { 00:16:57.380 "name": "pt1", 00:16:57.380 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:57.380 "is_configured": true, 00:16:57.380 "data_offset": 256, 00:16:57.380 "data_size": 7936 00:16:57.380 }, 00:16:57.380 { 00:16:57.380 "name": "pt2", 00:16:57.380 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:57.380 "is_configured": true, 00:16:57.380 "data_offset": 256, 00:16:57.380 "data_size": 7936 00:16:57.380 } 00:16:57.380 ] 00:16:57.380 } 00:16:57.380 } 00:16:57.380 }' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:57.380 pt2' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:57.380 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.381 [2024-10-29 11:06:02.847530] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:57.381 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=58571275-c117-4c19-a0b8-962295c41dda 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 58571275-c117-4c19-a0b8-962295c41dda ']' 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 [2024-10-29 11:06:02.895192] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:57.642 [2024-10-29 11:06:02.895221] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:57.642 [2024-10-29 11:06:02.895293] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:57.642 [2024-10-29 11:06:02.895352] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:57.642 [2024-10-29 11:06:02.895362] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:02 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 [2024-10-29 11:06:03.030980] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:57.642 [2024-10-29 11:06:03.032964] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:57.642 [2024-10-29 11:06:03.033094] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:57.642 [2024-10-29 11:06:03.033216] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:57.642 [2024-10-29 11:06:03.033285] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:57.642 [2024-10-29 11:06:03.033345] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state configuring 00:16:57.642 request: 00:16:57.642 { 00:16:57.642 "name": "raid_bdev1", 00:16:57.642 "raid_level": "raid1", 00:16:57.642 "base_bdevs": [ 00:16:57.642 "malloc1", 00:16:57.642 "malloc2" 00:16:57.642 ], 00:16:57.642 "superblock": false, 00:16:57.642 "method": "bdev_raid_create", 00:16:57.642 "req_id": 1 00:16:57.642 } 00:16:57.642 Got JSON-RPC error response 00:16:57.642 response: 00:16:57.642 { 00:16:57.642 "code": -17, 00:16:57.642 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:57.642 } 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.642 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.642 [2024-10-29 11:06:03.082841] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:57.642 [2024-10-29 11:06:03.082943] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:57.642 [2024-10-29 11:06:03.082980] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:57.643 [2024-10-29 11:06:03.083011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:57.643 [2024-10-29 11:06:03.084939] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:57.643 [2024-10-29 11:06:03.085032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:57.643 [2024-10-29 11:06:03.085101] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:57.643 [2024-10-29 11:06:03.085159] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:57.643 pt1 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:57.643 "name": "raid_bdev1", 00:16:57.643 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:57.643 "strip_size_kb": 0, 00:16:57.643 "state": "configuring", 00:16:57.643 "raid_level": "raid1", 00:16:57.643 "superblock": true, 00:16:57.643 "num_base_bdevs": 2, 00:16:57.643 "num_base_bdevs_discovered": 1, 00:16:57.643 "num_base_bdevs_operational": 2, 00:16:57.643 "base_bdevs_list": [ 00:16:57.643 { 00:16:57.643 "name": "pt1", 00:16:57.643 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:57.643 "is_configured": true, 00:16:57.643 "data_offset": 256, 00:16:57.643 "data_size": 7936 00:16:57.643 }, 00:16:57.643 { 00:16:57.643 "name": null, 00:16:57.643 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:57.643 "is_configured": false, 00:16:57.643 "data_offset": 256, 00:16:57.643 "data_size": 7936 00:16:57.643 } 00:16:57.643 ] 00:16:57.643 }' 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:57.643 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.213 [2024-10-29 11:06:03.530140] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:58.213 [2024-10-29 11:06:03.530211] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:58.213 [2024-10-29 11:06:03.530234] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:16:58.213 [2024-10-29 11:06:03.530244] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:58.213 [2024-10-29 11:06:03.530421] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:58.213 [2024-10-29 11:06:03.530437] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:58.213 [2024-10-29 11:06:03.530504] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:58.213 [2024-10-29 11:06:03.530536] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:58.213 [2024-10-29 11:06:03.530648] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006980 00:16:58.213 [2024-10-29 11:06:03.530660] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:58.213 [2024-10-29 11:06:03.530754] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:16:58.213 [2024-10-29 11:06:03.530855] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006980 00:16:58.213 [2024-10-29 11:06:03.530870] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006980 00:16:58.213 [2024-10-29 11:06:03.530932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:58.213 pt2 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:58.213 "name": "raid_bdev1", 00:16:58.213 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:58.213 "strip_size_kb": 0, 00:16:58.213 "state": "online", 00:16:58.213 "raid_level": "raid1", 00:16:58.213 "superblock": true, 00:16:58.213 "num_base_bdevs": 2, 00:16:58.213 "num_base_bdevs_discovered": 2, 00:16:58.213 "num_base_bdevs_operational": 2, 00:16:58.213 "base_bdevs_list": [ 00:16:58.213 { 00:16:58.213 "name": "pt1", 00:16:58.213 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:58.213 "is_configured": true, 00:16:58.213 "data_offset": 256, 00:16:58.213 "data_size": 7936 00:16:58.213 }, 00:16:58.213 { 00:16:58.213 "name": "pt2", 00:16:58.213 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:58.213 "is_configured": true, 00:16:58.213 "data_offset": 256, 00:16:58.213 "data_size": 7936 00:16:58.213 } 00:16:58.213 ] 00:16:58.213 }' 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:58.213 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.473 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:58.473 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:58.474 [2024-10-29 11:06:03.917728] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:58.474 "name": "raid_bdev1", 00:16:58.474 "aliases": [ 00:16:58.474 "58571275-c117-4c19-a0b8-962295c41dda" 00:16:58.474 ], 00:16:58.474 "product_name": "Raid Volume", 00:16:58.474 "block_size": 4128, 00:16:58.474 "num_blocks": 7936, 00:16:58.474 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:58.474 "md_size": 32, 00:16:58.474 "md_interleave": true, 00:16:58.474 "dif_type": 0, 00:16:58.474 "assigned_rate_limits": { 00:16:58.474 "rw_ios_per_sec": 0, 00:16:58.474 "rw_mbytes_per_sec": 0, 00:16:58.474 "r_mbytes_per_sec": 0, 00:16:58.474 "w_mbytes_per_sec": 0 00:16:58.474 }, 00:16:58.474 "claimed": false, 00:16:58.474 "zoned": false, 00:16:58.474 "supported_io_types": { 00:16:58.474 "read": true, 00:16:58.474 "write": true, 00:16:58.474 "unmap": false, 00:16:58.474 "flush": false, 00:16:58.474 "reset": true, 00:16:58.474 "nvme_admin": false, 00:16:58.474 "nvme_io": false, 00:16:58.474 "nvme_io_md": false, 00:16:58.474 "write_zeroes": true, 00:16:58.474 "zcopy": false, 00:16:58.474 "get_zone_info": false, 00:16:58.474 "zone_management": false, 00:16:58.474 "zone_append": false, 00:16:58.474 "compare": false, 00:16:58.474 "compare_and_write": false, 00:16:58.474 "abort": false, 00:16:58.474 "seek_hole": false, 00:16:58.474 "seek_data": false, 00:16:58.474 "copy": false, 00:16:58.474 "nvme_iov_md": false 00:16:58.474 }, 00:16:58.474 "memory_domains": [ 00:16:58.474 { 00:16:58.474 "dma_device_id": "system", 00:16:58.474 "dma_device_type": 1 00:16:58.474 }, 00:16:58.474 { 00:16:58.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:58.474 "dma_device_type": 2 00:16:58.474 }, 00:16:58.474 { 00:16:58.474 "dma_device_id": "system", 00:16:58.474 "dma_device_type": 1 00:16:58.474 }, 00:16:58.474 { 00:16:58.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:58.474 "dma_device_type": 2 00:16:58.474 } 00:16:58.474 ], 00:16:58.474 "driver_specific": { 00:16:58.474 "raid": { 00:16:58.474 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:58.474 "strip_size_kb": 0, 00:16:58.474 "state": "online", 00:16:58.474 "raid_level": "raid1", 00:16:58.474 "superblock": true, 00:16:58.474 "num_base_bdevs": 2, 00:16:58.474 "num_base_bdevs_discovered": 2, 00:16:58.474 "num_base_bdevs_operational": 2, 00:16:58.474 "base_bdevs_list": [ 00:16:58.474 { 00:16:58.474 "name": "pt1", 00:16:58.474 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:58.474 "is_configured": true, 00:16:58.474 "data_offset": 256, 00:16:58.474 "data_size": 7936 00:16:58.474 }, 00:16:58.474 { 00:16:58.474 "name": "pt2", 00:16:58.474 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:58.474 "is_configured": true, 00:16:58.474 "data_offset": 256, 00:16:58.474 "data_size": 7936 00:16:58.474 } 00:16:58.474 ] 00:16:58.474 } 00:16:58.474 } 00:16:58.474 }' 00:16:58.474 11:06:03 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:58.735 pt2' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:58.735 [2024-10-29 11:06:04.153347] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 58571275-c117-4c19-a0b8-962295c41dda '!=' 58571275-c117-4c19-a0b8-962295c41dda ']' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.735 [2024-10-29 11:06:04.181080] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.735 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.995 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:58.995 "name": "raid_bdev1", 00:16:58.995 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:58.995 "strip_size_kb": 0, 00:16:58.995 "state": "online", 00:16:58.995 "raid_level": "raid1", 00:16:58.995 "superblock": true, 00:16:58.995 "num_base_bdevs": 2, 00:16:58.995 "num_base_bdevs_discovered": 1, 00:16:58.995 "num_base_bdevs_operational": 1, 00:16:58.995 "base_bdevs_list": [ 00:16:58.995 { 00:16:58.995 "name": null, 00:16:58.995 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:58.995 "is_configured": false, 00:16:58.995 "data_offset": 0, 00:16:58.995 "data_size": 7936 00:16:58.995 }, 00:16:58.995 { 00:16:58.995 "name": "pt2", 00:16:58.995 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:58.995 "is_configured": true, 00:16:58.995 "data_offset": 256, 00:16:58.995 "data_size": 7936 00:16:58.995 } 00:16:58.995 ] 00:16:58.995 }' 00:16:58.995 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:58.995 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 [2024-10-29 11:06:04.624449] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:59.256 [2024-10-29 11:06:04.624536] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:59.256 [2024-10-29 11:06:04.624619] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:59.256 [2024-10-29 11:06:04.624729] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:59.256 [2024-10-29 11:06:04.624801] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006980 name raid_bdev1, state offline 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 [2024-10-29 11:06:04.696436] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:59.256 [2024-10-29 11:06:04.696487] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:59.256 [2024-10-29 11:06:04.696506] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:59.256 [2024-10-29 11:06:04.696516] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:59.256 [2024-10-29 11:06:04.698472] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:59.256 [2024-10-29 11:06:04.698513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:59.256 [2024-10-29 11:06:04.698563] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:59.256 [2024-10-29 11:06:04.698598] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:59.256 [2024-10-29 11:06:04.698659] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006d00 00:16:59.256 [2024-10-29 11:06:04.698668] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:59.256 [2024-10-29 11:06:04.698757] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:16:59.256 [2024-10-29 11:06:04.698821] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006d00 00:16:59.256 [2024-10-29 11:06:04.698832] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006d00 00:16:59.256 [2024-10-29 11:06:04.698896] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:59.256 pt2 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.256 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.516 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:59.516 "name": "raid_bdev1", 00:16:59.516 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:16:59.516 "strip_size_kb": 0, 00:16:59.516 "state": "online", 00:16:59.516 "raid_level": "raid1", 00:16:59.516 "superblock": true, 00:16:59.516 "num_base_bdevs": 2, 00:16:59.516 "num_base_bdevs_discovered": 1, 00:16:59.516 "num_base_bdevs_operational": 1, 00:16:59.516 "base_bdevs_list": [ 00:16:59.516 { 00:16:59.516 "name": null, 00:16:59.516 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:59.516 "is_configured": false, 00:16:59.516 "data_offset": 256, 00:16:59.516 "data_size": 7936 00:16:59.516 }, 00:16:59.516 { 00:16:59.516 "name": "pt2", 00:16:59.516 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:59.516 "is_configured": true, 00:16:59.516 "data_offset": 256, 00:16:59.516 "data_size": 7936 00:16:59.516 } 00:16:59.516 ] 00:16:59.516 }' 00:16:59.516 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:59.516 11:06:04 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.776 [2024-10-29 11:06:05.176046] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:59.776 [2024-10-29 11:06:05.176134] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:59.776 [2024-10-29 11:06:05.176225] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:59.776 [2024-10-29 11:06:05.176282] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:59.776 [2024-10-29 11:06:05.176376] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006d00 name raid_bdev1, state offline 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:59.776 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.777 [2024-10-29 11:06:05.239929] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:59.777 [2024-10-29 11:06:05.240038] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:59.777 [2024-10-29 11:06:05.240083] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:16:59.777 [2024-10-29 11:06:05.240123] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:59.777 [2024-10-29 11:06:05.242057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:59.777 [2024-10-29 11:06:05.242153] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:59.777 [2024-10-29 11:06:05.242223] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:59.777 [2024-10-29 11:06:05.242279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:59.777 [2024-10-29 11:06:05.242438] bdev_raid.c:3679:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:59.777 [2024-10-29 11:06:05.242504] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:59.777 [2024-10-29 11:06:05.242559] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007080 name raid_bdev1, state configuring 00:16:59.777 [2024-10-29 11:06:05.242641] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:59.777 [2024-10-29 11:06:05.242758] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007400 00:16:59.777 [2024-10-29 11:06:05.242805] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:59.777 [2024-10-29 11:06:05.242901] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:16:59.777 [2024-10-29 11:06:05.243007] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007400 00:16:59.777 [2024-10-29 11:06:05.243050] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007400 00:16:59.777 [2024-10-29 11:06:05.243189] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:59.777 pt1 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.777 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.037 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:00.037 "name": "raid_bdev1", 00:17:00.037 "uuid": "58571275-c117-4c19-a0b8-962295c41dda", 00:17:00.037 "strip_size_kb": 0, 00:17:00.037 "state": "online", 00:17:00.037 "raid_level": "raid1", 00:17:00.037 "superblock": true, 00:17:00.037 "num_base_bdevs": 2, 00:17:00.037 "num_base_bdevs_discovered": 1, 00:17:00.037 "num_base_bdevs_operational": 1, 00:17:00.037 "base_bdevs_list": [ 00:17:00.037 { 00:17:00.037 "name": null, 00:17:00.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:00.037 "is_configured": false, 00:17:00.037 "data_offset": 256, 00:17:00.037 "data_size": 7936 00:17:00.037 }, 00:17:00.037 { 00:17:00.037 "name": "pt2", 00:17:00.037 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:00.037 "is_configured": true, 00:17:00.037 "data_offset": 256, 00:17:00.037 "data_size": 7936 00:17:00.037 } 00:17:00.037 ] 00:17:00.037 }' 00:17:00.037 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:00.037 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.298 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.298 [2024-10-29 11:06:05.787220] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 58571275-c117-4c19-a0b8-962295c41dda '!=' 58571275-c117-4c19-a0b8-962295c41dda ']' 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 99131 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@952 -- # '[' -z 99131 ']' 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # kill -0 99131 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@957 -- # uname 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 99131 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@970 -- # echo 'killing process with pid 99131' 00:17:00.558 killing process with pid 99131 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@971 -- # kill 99131 00:17:00.558 [2024-10-29 11:06:05.860983] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:00.558 [2024-10-29 11:06:05.861127] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:00.558 11:06:05 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@976 -- # wait 99131 00:17:00.558 [2024-10-29 11:06:05.861219] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:00.558 [2024-10-29 11:06:05.861232] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007400 name raid_bdev1, state offline 00:17:00.558 [2024-10-29 11:06:05.885052] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:00.819 11:06:06 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:17:00.819 00:17:00.819 real 0m4.881s 00:17:00.819 user 0m7.957s 00:17:00.819 sys 0m1.089s 00:17:00.819 11:06:06 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:00.819 11:06:06 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.819 ************************************ 00:17:00.819 END TEST raid_superblock_test_md_interleaved 00:17:00.819 ************************************ 00:17:00.819 11:06:06 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:17:00.819 11:06:06 bdev_raid -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:17:00.819 11:06:06 bdev_raid -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:00.819 11:06:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:00.819 ************************************ 00:17:00.819 START TEST raid_rebuild_test_sb_md_interleaved 00:17:00.819 ************************************ 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1127 -- # raid_rebuild_test raid1 2 true false false 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=99445 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 99445 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@833 -- # '[' -z 99445 ']' 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # local max_retries=100 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:00.819 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # xtrace_disable 00:17:00.819 11:06:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.819 [2024-10-29 11:06:06.297899] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:00.819 [2024-10-29 11:06:06.298142] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:17:00.819 Zero copy mechanism will not be used. 00:17:00.819 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99445 ] 00:17:01.079 [2024-10-29 11:06:06.474315] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:01.080 [2024-10-29 11:06:06.499422] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:01.080 [2024-10-29 11:06:06.543653] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:01.080 [2024-10-29 11:06:06.543780] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@866 -- # return 0 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.650 BaseBdev1_malloc 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.650 [2024-10-29 11:06:07.135070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:01.650 [2024-10-29 11:06:07.135159] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:01.650 [2024-10-29 11:06:07.135192] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:01.650 [2024-10-29 11:06:07.135203] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:01.650 [2024-10-29 11:06:07.137190] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:01.650 [2024-10-29 11:06:07.137234] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:01.650 BaseBdev1 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.650 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 BaseBdev2_malloc 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 [2024-10-29 11:06:07.164000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:01.911 [2024-10-29 11:06:07.164062] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:01.911 [2024-10-29 11:06:07.164094] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:17:01.911 [2024-10-29 11:06:07.164104] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:01.911 [2024-10-29 11:06:07.166022] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:01.911 [2024-10-29 11:06:07.166064] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:01.911 BaseBdev2 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 spare_malloc 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 spare_delay 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 [2024-10-29 11:06:07.204927] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:01.911 [2024-10-29 11:06:07.204987] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:01.911 [2024-10-29 11:06:07.205010] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:17:01.911 [2024-10-29 11:06:07.205020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:01.911 [2024-10-29 11:06:07.206945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:01.911 [2024-10-29 11:06:07.206986] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:01.911 spare 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.911 [2024-10-29 11:06:07.216947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:01.911 [2024-10-29 11:06:07.218829] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:01.911 [2024-10-29 11:06:07.219010] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006280 00:17:01.911 [2024-10-29 11:06:07.219024] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:01.911 [2024-10-29 11:06:07.219112] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:17:01.911 [2024-10-29 11:06:07.219180] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006280 00:17:01.911 [2024-10-29 11:06:07.219192] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006280 00:17:01.911 [2024-10-29 11:06:07.219263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:01.911 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:01.912 "name": "raid_bdev1", 00:17:01.912 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:01.912 "strip_size_kb": 0, 00:17:01.912 "state": "online", 00:17:01.912 "raid_level": "raid1", 00:17:01.912 "superblock": true, 00:17:01.912 "num_base_bdevs": 2, 00:17:01.912 "num_base_bdevs_discovered": 2, 00:17:01.912 "num_base_bdevs_operational": 2, 00:17:01.912 "base_bdevs_list": [ 00:17:01.912 { 00:17:01.912 "name": "BaseBdev1", 00:17:01.912 "uuid": "d279fced-4b19-5f3a-8c9a-311f823b87bf", 00:17:01.912 "is_configured": true, 00:17:01.912 "data_offset": 256, 00:17:01.912 "data_size": 7936 00:17:01.912 }, 00:17:01.912 { 00:17:01.912 "name": "BaseBdev2", 00:17:01.912 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:01.912 "is_configured": true, 00:17:01.912 "data_offset": 256, 00:17:01.912 "data_size": 7936 00:17:01.912 } 00:17:01.912 ] 00:17:01.912 }' 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:01.912 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.483 [2024-10-29 11:06:07.680485] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.483 [2024-10-29 11:06:07.748182] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:02.483 "name": "raid_bdev1", 00:17:02.483 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:02.483 "strip_size_kb": 0, 00:17:02.483 "state": "online", 00:17:02.483 "raid_level": "raid1", 00:17:02.483 "superblock": true, 00:17:02.483 "num_base_bdevs": 2, 00:17:02.483 "num_base_bdevs_discovered": 1, 00:17:02.483 "num_base_bdevs_operational": 1, 00:17:02.483 "base_bdevs_list": [ 00:17:02.483 { 00:17:02.483 "name": null, 00:17:02.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:02.483 "is_configured": false, 00:17:02.483 "data_offset": 0, 00:17:02.483 "data_size": 7936 00:17:02.483 }, 00:17:02.483 { 00:17:02.483 "name": "BaseBdev2", 00:17:02.483 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:02.483 "is_configured": true, 00:17:02.483 "data_offset": 256, 00:17:02.483 "data_size": 7936 00:17:02.483 } 00:17:02.483 ] 00:17:02.483 }' 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:02.483 11:06:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.744 11:06:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:02.744 11:06:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.744 11:06:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.744 [2024-10-29 11:06:08.219837] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:02.744 [2024-10-29 11:06:08.234198] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:17:02.744 11:06:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.744 11:06:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:02.744 [2024-10-29 11:06:08.240724] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:04.129 "name": "raid_bdev1", 00:17:04.129 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:04.129 "strip_size_kb": 0, 00:17:04.129 "state": "online", 00:17:04.129 "raid_level": "raid1", 00:17:04.129 "superblock": true, 00:17:04.129 "num_base_bdevs": 2, 00:17:04.129 "num_base_bdevs_discovered": 2, 00:17:04.129 "num_base_bdevs_operational": 2, 00:17:04.129 "process": { 00:17:04.129 "type": "rebuild", 00:17:04.129 "target": "spare", 00:17:04.129 "progress": { 00:17:04.129 "blocks": 2560, 00:17:04.129 "percent": 32 00:17:04.129 } 00:17:04.129 }, 00:17:04.129 "base_bdevs_list": [ 00:17:04.129 { 00:17:04.129 "name": "spare", 00:17:04.129 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:04.129 "is_configured": true, 00:17:04.129 "data_offset": 256, 00:17:04.129 "data_size": 7936 00:17:04.129 }, 00:17:04.129 { 00:17:04.129 "name": "BaseBdev2", 00:17:04.129 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:04.129 "is_configured": true, 00:17:04.129 "data_offset": 256, 00:17:04.129 "data_size": 7936 00:17:04.129 } 00:17:04.129 ] 00:17:04.129 }' 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.129 [2024-10-29 11:06:09.396200] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:04.129 [2024-10-29 11:06:09.446607] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:04.129 [2024-10-29 11:06:09.446667] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:04.129 [2024-10-29 11:06:09.446686] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:04.129 [2024-10-29 11:06:09.446696] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.129 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:04.130 "name": "raid_bdev1", 00:17:04.130 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:04.130 "strip_size_kb": 0, 00:17:04.130 "state": "online", 00:17:04.130 "raid_level": "raid1", 00:17:04.130 "superblock": true, 00:17:04.130 "num_base_bdevs": 2, 00:17:04.130 "num_base_bdevs_discovered": 1, 00:17:04.130 "num_base_bdevs_operational": 1, 00:17:04.130 "base_bdevs_list": [ 00:17:04.130 { 00:17:04.130 "name": null, 00:17:04.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:04.130 "is_configured": false, 00:17:04.130 "data_offset": 0, 00:17:04.130 "data_size": 7936 00:17:04.130 }, 00:17:04.130 { 00:17:04.130 "name": "BaseBdev2", 00:17:04.130 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:04.130 "is_configured": true, 00:17:04.130 "data_offset": 256, 00:17:04.130 "data_size": 7936 00:17:04.130 } 00:17:04.130 ] 00:17:04.130 }' 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:04.130 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:04.698 "name": "raid_bdev1", 00:17:04.698 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:04.698 "strip_size_kb": 0, 00:17:04.698 "state": "online", 00:17:04.698 "raid_level": "raid1", 00:17:04.698 "superblock": true, 00:17:04.698 "num_base_bdevs": 2, 00:17:04.698 "num_base_bdevs_discovered": 1, 00:17:04.698 "num_base_bdevs_operational": 1, 00:17:04.698 "base_bdevs_list": [ 00:17:04.698 { 00:17:04.698 "name": null, 00:17:04.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:04.698 "is_configured": false, 00:17:04.698 "data_offset": 0, 00:17:04.698 "data_size": 7936 00:17:04.698 }, 00:17:04.698 { 00:17:04.698 "name": "BaseBdev2", 00:17:04.698 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:04.698 "is_configured": true, 00:17:04.698 "data_offset": 256, 00:17:04.698 "data_size": 7936 00:17:04.698 } 00:17:04.698 ] 00:17:04.698 }' 00:17:04.698 11:06:09 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.698 [2024-10-29 11:06:10.054104] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:04.698 [2024-10-29 11:06:10.057202] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:17:04.698 [2024-10-29 11:06:10.059121] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.698 11:06:10 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:05.637 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:05.637 "name": "raid_bdev1", 00:17:05.637 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:05.637 "strip_size_kb": 0, 00:17:05.637 "state": "online", 00:17:05.637 "raid_level": "raid1", 00:17:05.637 "superblock": true, 00:17:05.637 "num_base_bdevs": 2, 00:17:05.637 "num_base_bdevs_discovered": 2, 00:17:05.637 "num_base_bdevs_operational": 2, 00:17:05.637 "process": { 00:17:05.637 "type": "rebuild", 00:17:05.637 "target": "spare", 00:17:05.637 "progress": { 00:17:05.637 "blocks": 2560, 00:17:05.637 "percent": 32 00:17:05.637 } 00:17:05.637 }, 00:17:05.637 "base_bdevs_list": [ 00:17:05.637 { 00:17:05.637 "name": "spare", 00:17:05.637 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:05.637 "is_configured": true, 00:17:05.637 "data_offset": 256, 00:17:05.637 "data_size": 7936 00:17:05.637 }, 00:17:05.637 { 00:17:05.637 "name": "BaseBdev2", 00:17:05.637 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:05.637 "is_configured": true, 00:17:05.637 "data_offset": 256, 00:17:05.637 "data_size": 7936 00:17:05.637 } 00:17:05.638 ] 00:17:05.638 }' 00:17:05.638 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:05.898 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=616 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:05.898 "name": "raid_bdev1", 00:17:05.898 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:05.898 "strip_size_kb": 0, 00:17:05.898 "state": "online", 00:17:05.898 "raid_level": "raid1", 00:17:05.898 "superblock": true, 00:17:05.898 "num_base_bdevs": 2, 00:17:05.898 "num_base_bdevs_discovered": 2, 00:17:05.898 "num_base_bdevs_operational": 2, 00:17:05.898 "process": { 00:17:05.898 "type": "rebuild", 00:17:05.898 "target": "spare", 00:17:05.898 "progress": { 00:17:05.898 "blocks": 2816, 00:17:05.898 "percent": 35 00:17:05.898 } 00:17:05.898 }, 00:17:05.898 "base_bdevs_list": [ 00:17:05.898 { 00:17:05.898 "name": "spare", 00:17:05.898 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:05.898 "is_configured": true, 00:17:05.898 "data_offset": 256, 00:17:05.898 "data_size": 7936 00:17:05.898 }, 00:17:05.898 { 00:17:05.898 "name": "BaseBdev2", 00:17:05.898 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:05.898 "is_configured": true, 00:17:05.898 "data_offset": 256, 00:17:05.898 "data_size": 7936 00:17:05.898 } 00:17:05.898 ] 00:17:05.898 }' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:05.898 11:06:11 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.839 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:07.112 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:07.112 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.112 "name": "raid_bdev1", 00:17:07.112 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:07.112 "strip_size_kb": 0, 00:17:07.113 "state": "online", 00:17:07.113 "raid_level": "raid1", 00:17:07.113 "superblock": true, 00:17:07.113 "num_base_bdevs": 2, 00:17:07.113 "num_base_bdevs_discovered": 2, 00:17:07.113 "num_base_bdevs_operational": 2, 00:17:07.113 "process": { 00:17:07.113 "type": "rebuild", 00:17:07.113 "target": "spare", 00:17:07.113 "progress": { 00:17:07.113 "blocks": 5632, 00:17:07.113 "percent": 70 00:17:07.113 } 00:17:07.113 }, 00:17:07.113 "base_bdevs_list": [ 00:17:07.113 { 00:17:07.113 "name": "spare", 00:17:07.113 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:07.113 "is_configured": true, 00:17:07.113 "data_offset": 256, 00:17:07.113 "data_size": 7936 00:17:07.113 }, 00:17:07.113 { 00:17:07.113 "name": "BaseBdev2", 00:17:07.113 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:07.113 "is_configured": true, 00:17:07.113 "data_offset": 256, 00:17:07.113 "data_size": 7936 00:17:07.113 } 00:17:07.113 ] 00:17:07.113 }' 00:17:07.113 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.113 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:07.113 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.113 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:07.113 11:06:12 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:07.751 [2024-10-29 11:06:13.170139] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:07.751 [2024-10-29 11:06:13.170235] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:07.751 [2024-10-29 11:06:13.170338] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.011 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:08.271 "name": "raid_bdev1", 00:17:08.271 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:08.271 "strip_size_kb": 0, 00:17:08.271 "state": "online", 00:17:08.271 "raid_level": "raid1", 00:17:08.271 "superblock": true, 00:17:08.271 "num_base_bdevs": 2, 00:17:08.271 "num_base_bdevs_discovered": 2, 00:17:08.271 "num_base_bdevs_operational": 2, 00:17:08.271 "base_bdevs_list": [ 00:17:08.271 { 00:17:08.271 "name": "spare", 00:17:08.271 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:08.271 "is_configured": true, 00:17:08.271 "data_offset": 256, 00:17:08.271 "data_size": 7936 00:17:08.271 }, 00:17:08.271 { 00:17:08.271 "name": "BaseBdev2", 00:17:08.271 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:08.271 "is_configured": true, 00:17:08.271 "data_offset": 256, 00:17:08.271 "data_size": 7936 00:17:08.271 } 00:17:08.271 ] 00:17:08.271 }' 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.271 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:08.271 "name": "raid_bdev1", 00:17:08.271 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:08.271 "strip_size_kb": 0, 00:17:08.271 "state": "online", 00:17:08.271 "raid_level": "raid1", 00:17:08.271 "superblock": true, 00:17:08.271 "num_base_bdevs": 2, 00:17:08.271 "num_base_bdevs_discovered": 2, 00:17:08.271 "num_base_bdevs_operational": 2, 00:17:08.271 "base_bdevs_list": [ 00:17:08.271 { 00:17:08.271 "name": "spare", 00:17:08.271 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:08.271 "is_configured": true, 00:17:08.271 "data_offset": 256, 00:17:08.271 "data_size": 7936 00:17:08.271 }, 00:17:08.271 { 00:17:08.271 "name": "BaseBdev2", 00:17:08.271 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:08.272 "is_configured": true, 00:17:08.272 "data_offset": 256, 00:17:08.272 "data_size": 7936 00:17:08.272 } 00:17:08.272 ] 00:17:08.272 }' 00:17:08.272 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:08.272 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:08.272 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:08.532 "name": "raid_bdev1", 00:17:08.532 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:08.532 "strip_size_kb": 0, 00:17:08.532 "state": "online", 00:17:08.532 "raid_level": "raid1", 00:17:08.532 "superblock": true, 00:17:08.532 "num_base_bdevs": 2, 00:17:08.532 "num_base_bdevs_discovered": 2, 00:17:08.532 "num_base_bdevs_operational": 2, 00:17:08.532 "base_bdevs_list": [ 00:17:08.532 { 00:17:08.532 "name": "spare", 00:17:08.532 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:08.532 "is_configured": true, 00:17:08.532 "data_offset": 256, 00:17:08.532 "data_size": 7936 00:17:08.532 }, 00:17:08.532 { 00:17:08.532 "name": "BaseBdev2", 00:17:08.532 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:08.532 "is_configured": true, 00:17:08.532 "data_offset": 256, 00:17:08.532 "data_size": 7936 00:17:08.532 } 00:17:08.532 ] 00:17:08.532 }' 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:08.532 11:06:13 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.792 [2024-10-29 11:06:14.276474] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:08.792 [2024-10-29 11:06:14.276563] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:08.792 [2024-10-29 11:06:14.276675] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:08.792 [2024-10-29 11:06:14.276818] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:08.792 [2024-10-29 11:06:14.276882] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006280 name raid_bdev1, state offline 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.792 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.052 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.052 [2024-10-29 11:06:14.348430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:09.052 [2024-10-29 11:06:14.348490] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:09.052 [2024-10-29 11:06:14.348511] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:17:09.052 [2024-10-29 11:06:14.348525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:09.052 [2024-10-29 11:06:14.350474] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:09.052 [2024-10-29 11:06:14.350523] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:09.052 [2024-10-29 11:06:14.350579] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:09.052 [2024-10-29 11:06:14.350634] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:09.053 [2024-10-29 11:06:14.350767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:09.053 spare 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.053 [2024-10-29 11:06:14.450660] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000006600 00:17:09.053 [2024-10-29 11:06:14.450732] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:09.053 [2024-10-29 11:06:14.450839] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:17:09.053 [2024-10-29 11:06:14.450983] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000006600 00:17:09.053 [2024-10-29 11:06:14.451029] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000006600 00:17:09.053 [2024-10-29 11:06:14.451193] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:09.053 "name": "raid_bdev1", 00:17:09.053 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:09.053 "strip_size_kb": 0, 00:17:09.053 "state": "online", 00:17:09.053 "raid_level": "raid1", 00:17:09.053 "superblock": true, 00:17:09.053 "num_base_bdevs": 2, 00:17:09.053 "num_base_bdevs_discovered": 2, 00:17:09.053 "num_base_bdevs_operational": 2, 00:17:09.053 "base_bdevs_list": [ 00:17:09.053 { 00:17:09.053 "name": "spare", 00:17:09.053 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:09.053 "is_configured": true, 00:17:09.053 "data_offset": 256, 00:17:09.053 "data_size": 7936 00:17:09.053 }, 00:17:09.053 { 00:17:09.053 "name": "BaseBdev2", 00:17:09.053 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:09.053 "is_configured": true, 00:17:09.053 "data_offset": 256, 00:17:09.053 "data_size": 7936 00:17:09.053 } 00:17:09.053 ] 00:17:09.053 }' 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:09.053 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:09.623 "name": "raid_bdev1", 00:17:09.623 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:09.623 "strip_size_kb": 0, 00:17:09.623 "state": "online", 00:17:09.623 "raid_level": "raid1", 00:17:09.623 "superblock": true, 00:17:09.623 "num_base_bdevs": 2, 00:17:09.623 "num_base_bdevs_discovered": 2, 00:17:09.623 "num_base_bdevs_operational": 2, 00:17:09.623 "base_bdevs_list": [ 00:17:09.623 { 00:17:09.623 "name": "spare", 00:17:09.623 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:09.623 "is_configured": true, 00:17:09.623 "data_offset": 256, 00:17:09.623 "data_size": 7936 00:17:09.623 }, 00:17:09.623 { 00:17:09.623 "name": "BaseBdev2", 00:17:09.623 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:09.623 "is_configured": true, 00:17:09.623 "data_offset": 256, 00:17:09.623 "data_size": 7936 00:17:09.623 } 00:17:09.623 ] 00:17:09.623 }' 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:09.623 11:06:14 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:09.623 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:09.623 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:09.623 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.623 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.623 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.624 [2024-10-29 11:06:15.071864] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:09.624 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:09.884 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:09.884 "name": "raid_bdev1", 00:17:09.884 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:09.884 "strip_size_kb": 0, 00:17:09.884 "state": "online", 00:17:09.884 "raid_level": "raid1", 00:17:09.884 "superblock": true, 00:17:09.884 "num_base_bdevs": 2, 00:17:09.884 "num_base_bdevs_discovered": 1, 00:17:09.884 "num_base_bdevs_operational": 1, 00:17:09.884 "base_bdevs_list": [ 00:17:09.884 { 00:17:09.884 "name": null, 00:17:09.884 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:09.884 "is_configured": false, 00:17:09.884 "data_offset": 0, 00:17:09.884 "data_size": 7936 00:17:09.884 }, 00:17:09.884 { 00:17:09.884 "name": "BaseBdev2", 00:17:09.884 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:09.884 "is_configured": true, 00:17:09.884 "data_offset": 256, 00:17:09.884 "data_size": 7936 00:17:09.884 } 00:17:09.884 ] 00:17:09.884 }' 00:17:09.884 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:09.884 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.143 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:10.143 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.143 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.143 [2024-10-29 11:06:15.515149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:10.143 [2024-10-29 11:06:15.515412] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:10.143 [2024-10-29 11:06:15.515482] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:10.143 [2024-10-29 11:06:15.515558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:10.143 [2024-10-29 11:06:15.519315] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:17:10.143 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.143 11:06:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:10.143 [2024-10-29 11:06:15.521294] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:11.083 "name": "raid_bdev1", 00:17:11.083 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:11.083 "strip_size_kb": 0, 00:17:11.083 "state": "online", 00:17:11.083 "raid_level": "raid1", 00:17:11.083 "superblock": true, 00:17:11.083 "num_base_bdevs": 2, 00:17:11.083 "num_base_bdevs_discovered": 2, 00:17:11.083 "num_base_bdevs_operational": 2, 00:17:11.083 "process": { 00:17:11.083 "type": "rebuild", 00:17:11.083 "target": "spare", 00:17:11.083 "progress": { 00:17:11.083 "blocks": 2560, 00:17:11.083 "percent": 32 00:17:11.083 } 00:17:11.083 }, 00:17:11.083 "base_bdevs_list": [ 00:17:11.083 { 00:17:11.083 "name": "spare", 00:17:11.083 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:11.083 "is_configured": true, 00:17:11.083 "data_offset": 256, 00:17:11.083 "data_size": 7936 00:17:11.083 }, 00:17:11.083 { 00:17:11.083 "name": "BaseBdev2", 00:17:11.083 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:11.083 "is_configured": true, 00:17:11.083 "data_offset": 256, 00:17:11.083 "data_size": 7936 00:17:11.083 } 00:17:11.083 ] 00:17:11.083 }' 00:17:11.083 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.343 [2024-10-29 11:06:16.681941] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:11.343 [2024-10-29 11:06:16.725464] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:11.343 [2024-10-29 11:06:16.725572] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:11.343 [2024-10-29 11:06:16.725614] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:11.343 [2024-10-29 11:06:16.725653] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:11.343 "name": "raid_bdev1", 00:17:11.343 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:11.343 "strip_size_kb": 0, 00:17:11.343 "state": "online", 00:17:11.343 "raid_level": "raid1", 00:17:11.343 "superblock": true, 00:17:11.343 "num_base_bdevs": 2, 00:17:11.343 "num_base_bdevs_discovered": 1, 00:17:11.343 "num_base_bdevs_operational": 1, 00:17:11.343 "base_bdevs_list": [ 00:17:11.343 { 00:17:11.343 "name": null, 00:17:11.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:11.343 "is_configured": false, 00:17:11.343 "data_offset": 0, 00:17:11.343 "data_size": 7936 00:17:11.343 }, 00:17:11.343 { 00:17:11.343 "name": "BaseBdev2", 00:17:11.343 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:11.343 "is_configured": true, 00:17:11.343 "data_offset": 256, 00:17:11.343 "data_size": 7936 00:17:11.343 } 00:17:11.343 ] 00:17:11.343 }' 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:11.343 11:06:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.914 11:06:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:11.914 11:06:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.914 11:06:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.914 [2024-10-29 11:06:17.184739] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:11.914 [2024-10-29 11:06:17.184809] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:11.914 [2024-10-29 11:06:17.184838] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:17:11.914 [2024-10-29 11:06:17.184848] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:11.914 [2024-10-29 11:06:17.185050] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:11.914 [2024-10-29 11:06:17.185066] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:11.914 [2024-10-29 11:06:17.185129] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:11.914 [2024-10-29 11:06:17.185141] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:11.914 [2024-10-29 11:06:17.185154] bdev_raid.c:3752:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:11.914 [2024-10-29 11:06:17.185180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:11.914 [2024-10-29 11:06:17.188841] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:17:11.914 spare 00:17:11.914 11:06:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.914 11:06:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:11.914 [2024-10-29 11:06:17.190734] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:12.854 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:12.854 "name": "raid_bdev1", 00:17:12.854 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:12.854 "strip_size_kb": 0, 00:17:12.854 "state": "online", 00:17:12.854 "raid_level": "raid1", 00:17:12.854 "superblock": true, 00:17:12.855 "num_base_bdevs": 2, 00:17:12.855 "num_base_bdevs_discovered": 2, 00:17:12.855 "num_base_bdevs_operational": 2, 00:17:12.855 "process": { 00:17:12.855 "type": "rebuild", 00:17:12.855 "target": "spare", 00:17:12.855 "progress": { 00:17:12.855 "blocks": 2560, 00:17:12.855 "percent": 32 00:17:12.855 } 00:17:12.855 }, 00:17:12.855 "base_bdevs_list": [ 00:17:12.855 { 00:17:12.855 "name": "spare", 00:17:12.855 "uuid": "5f96cac1-dd8f-554d-89a9-847898c81efe", 00:17:12.855 "is_configured": true, 00:17:12.855 "data_offset": 256, 00:17:12.855 "data_size": 7936 00:17:12.855 }, 00:17:12.855 { 00:17:12.855 "name": "BaseBdev2", 00:17:12.855 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:12.855 "is_configured": true, 00:17:12.855 "data_offset": 256, 00:17:12.855 "data_size": 7936 00:17:12.855 } 00:17:12.855 ] 00:17:12.855 }' 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:12.855 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:12.855 [2024-10-29 11:06:18.351787] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:13.115 [2024-10-29 11:06:18.394799] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:13.115 [2024-10-29 11:06:18.394874] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:13.115 [2024-10-29 11:06:18.394890] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:13.115 [2024-10-29 11:06:18.394902] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:13.115 "name": "raid_bdev1", 00:17:13.115 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:13.115 "strip_size_kb": 0, 00:17:13.115 "state": "online", 00:17:13.115 "raid_level": "raid1", 00:17:13.115 "superblock": true, 00:17:13.115 "num_base_bdevs": 2, 00:17:13.115 "num_base_bdevs_discovered": 1, 00:17:13.115 "num_base_bdevs_operational": 1, 00:17:13.115 "base_bdevs_list": [ 00:17:13.115 { 00:17:13.115 "name": null, 00:17:13.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:13.115 "is_configured": false, 00:17:13.115 "data_offset": 0, 00:17:13.115 "data_size": 7936 00:17:13.115 }, 00:17:13.115 { 00:17:13.115 "name": "BaseBdev2", 00:17:13.115 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:13.115 "is_configured": true, 00:17:13.115 "data_offset": 256, 00:17:13.115 "data_size": 7936 00:17:13.115 } 00:17:13.115 ] 00:17:13.115 }' 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:13.115 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.376 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.636 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:13.636 "name": "raid_bdev1", 00:17:13.636 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:13.636 "strip_size_kb": 0, 00:17:13.636 "state": "online", 00:17:13.636 "raid_level": "raid1", 00:17:13.636 "superblock": true, 00:17:13.636 "num_base_bdevs": 2, 00:17:13.636 "num_base_bdevs_discovered": 1, 00:17:13.636 "num_base_bdevs_operational": 1, 00:17:13.636 "base_bdevs_list": [ 00:17:13.636 { 00:17:13.636 "name": null, 00:17:13.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:13.636 "is_configured": false, 00:17:13.636 "data_offset": 0, 00:17:13.636 "data_size": 7936 00:17:13.636 }, 00:17:13.636 { 00:17:13.636 "name": "BaseBdev2", 00:17:13.636 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:13.636 "is_configured": true, 00:17:13.636 "data_offset": 256, 00:17:13.636 "data_size": 7936 00:17:13.636 } 00:17:13.636 ] 00:17:13.636 }' 00:17:13.636 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:13.636 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:13.636 11:06:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:13.636 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.637 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.637 [2024-10-29 11:06:19.021900] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:13.637 [2024-10-29 11:06:19.021966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:13.637 [2024-10-29 11:06:19.021987] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:17:13.637 [2024-10-29 11:06:19.021999] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:13.637 [2024-10-29 11:06:19.022145] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:13.637 [2024-10-29 11:06:19.022162] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:13.637 [2024-10-29 11:06:19.022209] bdev_raid.c:3901:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:13.637 [2024-10-29 11:06:19.022237] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:13.637 [2024-10-29 11:06:19.022246] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:13.637 [2024-10-29 11:06:19.022261] bdev_raid.c:3888:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:13.637 BaseBdev1 00:17:13.637 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.637 11:06:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:14.577 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:14.837 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:14.838 "name": "raid_bdev1", 00:17:14.838 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:14.838 "strip_size_kb": 0, 00:17:14.838 "state": "online", 00:17:14.838 "raid_level": "raid1", 00:17:14.838 "superblock": true, 00:17:14.838 "num_base_bdevs": 2, 00:17:14.838 "num_base_bdevs_discovered": 1, 00:17:14.838 "num_base_bdevs_operational": 1, 00:17:14.838 "base_bdevs_list": [ 00:17:14.838 { 00:17:14.838 "name": null, 00:17:14.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:14.838 "is_configured": false, 00:17:14.838 "data_offset": 0, 00:17:14.838 "data_size": 7936 00:17:14.838 }, 00:17:14.838 { 00:17:14.838 "name": "BaseBdev2", 00:17:14.838 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:14.838 "is_configured": true, 00:17:14.838 "data_offset": 256, 00:17:14.838 "data_size": 7936 00:17:14.838 } 00:17:14.838 ] 00:17:14.838 }' 00:17:14.838 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:14.838 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:15.099 "name": "raid_bdev1", 00:17:15.099 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:15.099 "strip_size_kb": 0, 00:17:15.099 "state": "online", 00:17:15.099 "raid_level": "raid1", 00:17:15.099 "superblock": true, 00:17:15.099 "num_base_bdevs": 2, 00:17:15.099 "num_base_bdevs_discovered": 1, 00:17:15.099 "num_base_bdevs_operational": 1, 00:17:15.099 "base_bdevs_list": [ 00:17:15.099 { 00:17:15.099 "name": null, 00:17:15.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:15.099 "is_configured": false, 00:17:15.099 "data_offset": 0, 00:17:15.099 "data_size": 7936 00:17:15.099 }, 00:17:15.099 { 00:17:15.099 "name": "BaseBdev2", 00:17:15.099 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:15.099 "is_configured": true, 00:17:15.099 "data_offset": 256, 00:17:15.099 "data_size": 7936 00:17:15.099 } 00:17:15.099 ] 00:17:15.099 }' 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:15.099 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.360 [2024-10-29 11:06:20.635270] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:15.360 [2024-10-29 11:06:20.635522] bdev_raid.c:3694:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:15.360 [2024-10-29 11:06:20.635598] bdev_raid.c:3713:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:15.360 request: 00:17:15.360 { 00:17:15.360 "base_bdev": "BaseBdev1", 00:17:15.360 "raid_bdev": "raid_bdev1", 00:17:15.360 "method": "bdev_raid_add_base_bdev", 00:17:15.360 "req_id": 1 00:17:15.360 } 00:17:15.360 Got JSON-RPC error response 00:17:15.360 response: 00:17:15.360 { 00:17:15.360 "code": -22, 00:17:15.360 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:15.360 } 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:15.360 11:06:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:16.301 "name": "raid_bdev1", 00:17:16.301 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:16.301 "strip_size_kb": 0, 00:17:16.301 "state": "online", 00:17:16.301 "raid_level": "raid1", 00:17:16.301 "superblock": true, 00:17:16.301 "num_base_bdevs": 2, 00:17:16.301 "num_base_bdevs_discovered": 1, 00:17:16.301 "num_base_bdevs_operational": 1, 00:17:16.301 "base_bdevs_list": [ 00:17:16.301 { 00:17:16.301 "name": null, 00:17:16.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:16.301 "is_configured": false, 00:17:16.301 "data_offset": 0, 00:17:16.301 "data_size": 7936 00:17:16.301 }, 00:17:16.301 { 00:17:16.301 "name": "BaseBdev2", 00:17:16.301 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:16.301 "is_configured": true, 00:17:16.301 "data_offset": 256, 00:17:16.301 "data_size": 7936 00:17:16.301 } 00:17:16.301 ] 00:17:16.301 }' 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:16.301 11:06:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:16.872 "name": "raid_bdev1", 00:17:16.872 "uuid": "f54a4a4c-fc88-4b36-a84c-c4bef796159f", 00:17:16.872 "strip_size_kb": 0, 00:17:16.872 "state": "online", 00:17:16.872 "raid_level": "raid1", 00:17:16.872 "superblock": true, 00:17:16.872 "num_base_bdevs": 2, 00:17:16.872 "num_base_bdevs_discovered": 1, 00:17:16.872 "num_base_bdevs_operational": 1, 00:17:16.872 "base_bdevs_list": [ 00:17:16.872 { 00:17:16.872 "name": null, 00:17:16.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:16.872 "is_configured": false, 00:17:16.872 "data_offset": 0, 00:17:16.872 "data_size": 7936 00:17:16.872 }, 00:17:16.872 { 00:17:16.872 "name": "BaseBdev2", 00:17:16.872 "uuid": "fae6e6a3-c805-5dbb-838e-473e47adee4d", 00:17:16.872 "is_configured": true, 00:17:16.872 "data_offset": 256, 00:17:16.872 "data_size": 7936 00:17:16.872 } 00:17:16.872 ] 00:17:16.872 }' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 99445 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@952 -- # '[' -z 99445 ']' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # kill -0 99445 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@957 -- # uname 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 99445 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:16.872 killing process with pid 99445 00:17:16.872 Received shutdown signal, test time was about 60.000000 seconds 00:17:16.872 00:17:16.872 Latency(us) 00:17:16.872 [2024-10-29T11:06:22.370Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:16.872 [2024-10-29T11:06:22.370Z] =================================================================================================================== 00:17:16.872 [2024-10-29T11:06:22.370Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@970 -- # echo 'killing process with pid 99445' 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@971 -- # kill 99445 00:17:16.872 [2024-10-29 11:06:22.314189] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:16.872 [2024-10-29 11:06:22.314316] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:16.872 [2024-10-29 11:06:22.314389] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:16.872 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@976 -- # wait 99445 00:17:16.872 [2024-10-29 11:06:22.314399] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000006600 name raid_bdev1, state offline 00:17:16.872 [2024-10-29 11:06:22.347624] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:17.133 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:17:17.133 00:17:17.133 real 0m16.361s 00:17:17.133 user 0m21.922s 00:17:17.133 sys 0m1.753s 00:17:17.133 ************************************ 00:17:17.133 END TEST raid_rebuild_test_sb_md_interleaved 00:17:17.133 ************************************ 00:17:17.133 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:17.133 11:06:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:17.133 11:06:22 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:17:17.133 11:06:22 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:17:17.133 11:06:22 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 99445 ']' 00:17:17.133 11:06:22 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 99445 00:17:17.394 11:06:22 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:17:17.394 00:17:17.394 real 9m56.728s 00:17:17.394 user 14m4.549s 00:17:17.394 sys 1m49.937s 00:17:17.394 11:06:22 bdev_raid -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:17.394 ************************************ 00:17:17.394 END TEST bdev_raid 00:17:17.394 ************************************ 00:17:17.394 11:06:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:17.394 11:06:22 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:17.394 11:06:22 -- common/autotest_common.sh@1103 -- # '[' 2 -le 1 ']' 00:17:17.394 11:06:22 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:17.394 11:06:22 -- common/autotest_common.sh@10 -- # set +x 00:17:17.394 ************************************ 00:17:17.394 START TEST spdkcli_raid 00:17:17.394 ************************************ 00:17:17.394 11:06:22 spdkcli_raid -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:17.394 * Looking for test storage... 00:17:17.394 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:17.394 11:06:22 spdkcli_raid -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:17:17.394 11:06:22 spdkcli_raid -- common/autotest_common.sh@1691 -- # lcov --version 00:17:17.394 11:06:22 spdkcli_raid -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:17.655 11:06:22 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:17:17.655 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:17.655 --rc genhtml_branch_coverage=1 00:17:17.655 --rc genhtml_function_coverage=1 00:17:17.655 --rc genhtml_legend=1 00:17:17.655 --rc geninfo_all_blocks=1 00:17:17.655 --rc geninfo_unexecuted_blocks=1 00:17:17.655 00:17:17.655 ' 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:17:17.655 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:17.655 --rc genhtml_branch_coverage=1 00:17:17.655 --rc genhtml_function_coverage=1 00:17:17.655 --rc genhtml_legend=1 00:17:17.655 --rc geninfo_all_blocks=1 00:17:17.655 --rc geninfo_unexecuted_blocks=1 00:17:17.655 00:17:17.655 ' 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:17:17.655 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:17.655 --rc genhtml_branch_coverage=1 00:17:17.655 --rc genhtml_function_coverage=1 00:17:17.655 --rc genhtml_legend=1 00:17:17.655 --rc geninfo_all_blocks=1 00:17:17.655 --rc geninfo_unexecuted_blocks=1 00:17:17.655 00:17:17.655 ' 00:17:17.655 11:06:22 spdkcli_raid -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:17:17.655 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:17.655 --rc genhtml_branch_coverage=1 00:17:17.655 --rc genhtml_function_coverage=1 00:17:17.655 --rc genhtml_legend=1 00:17:17.655 --rc geninfo_all_blocks=1 00:17:17.655 --rc geninfo_unexecuted_blocks=1 00:17:17.655 00:17:17.655 ' 00:17:17.655 11:06:22 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:17.655 11:06:22 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:17.655 11:06:22 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:17.655 11:06:22 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:17:17.655 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:17:17.655 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:17:17.655 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:17:17.655 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:17:17.656 11:06:22 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:17:17.656 11:06:22 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:17:17.656 11:06:22 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:17.656 11:06:22 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:17.656 11:06:23 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:17:17.656 11:06:23 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=100117 00:17:17.656 11:06:23 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:17:17.656 11:06:23 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 100117 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@833 -- # '[' -z 100117 ']' 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@838 -- # local max_retries=100 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:17.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@842 -- # xtrace_disable 00:17:17.656 11:06:23 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:17.656 [2024-10-29 11:06:23.099161] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:17.656 [2024-10-29 11:06:23.099354] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100117 ] 00:17:17.915 [2024-10-29 11:06:23.273309] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:17.915 [2024-10-29 11:06:23.301454] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:17.915 [2024-10-29 11:06:23.301549] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@866 -- # return 0 00:17:18.486 11:06:23 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:18.486 11:06:23 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:18.486 11:06:23 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:18.486 11:06:23 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:17:18.486 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:17:18.486 ' 00:17:20.398 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:17:20.398 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:17:20.398 11:06:25 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:17:20.398 11:06:25 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:20.398 11:06:25 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:20.398 11:06:25 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:17:20.398 11:06:25 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:20.398 11:06:25 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:20.398 11:06:25 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:17:20.398 ' 00:17:21.339 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:17:21.340 11:06:26 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:17:21.340 11:06:26 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:21.340 11:06:26 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:21.600 11:06:26 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:17:21.600 11:06:26 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:21.600 11:06:26 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:21.600 11:06:26 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:17:21.600 11:06:26 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:17:22.170 11:06:27 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:17:22.170 11:06:27 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:17:22.170 11:06:27 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:17:22.170 11:06:27 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:22.170 11:06:27 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:22.170 11:06:27 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:17:22.170 11:06:27 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:22.170 11:06:27 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:22.170 11:06:27 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:17:22.170 ' 00:17:23.111 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:17:23.111 11:06:28 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:17:23.111 11:06:28 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:23.111 11:06:28 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.372 11:06:28 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:17:23.372 11:06:28 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:23.372 11:06:28 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.372 11:06:28 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:17:23.372 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:17:23.372 ' 00:17:24.761 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:17:24.761 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:17:24.761 11:06:30 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:24.761 11:06:30 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 100117 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@952 -- # '[' -z 100117 ']' 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@956 -- # kill -0 100117 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@957 -- # uname 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 100117 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@970 -- # echo 'killing process with pid 100117' 00:17:24.761 killing process with pid 100117 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@971 -- # kill 100117 00:17:24.761 11:06:30 spdkcli_raid -- common/autotest_common.sh@976 -- # wait 100117 00:17:25.021 11:06:30 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 100117 ']' 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 100117 00:17:25.281 11:06:30 spdkcli_raid -- common/autotest_common.sh@952 -- # '[' -z 100117 ']' 00:17:25.281 11:06:30 spdkcli_raid -- common/autotest_common.sh@956 -- # kill -0 100117 00:17:25.281 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 956: kill: (100117) - No such process 00:17:25.281 11:06:30 spdkcli_raid -- common/autotest_common.sh@979 -- # echo 'Process with pid 100117 is not found' 00:17:25.281 Process with pid 100117 is not found 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:17:25.281 11:06:30 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:17:25.281 00:17:25.281 real 0m7.798s 00:17:25.281 user 0m16.475s 00:17:25.281 sys 0m1.157s 00:17:25.281 11:06:30 spdkcli_raid -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:25.281 11:06:30 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:25.281 ************************************ 00:17:25.281 END TEST spdkcli_raid 00:17:25.281 ************************************ 00:17:25.281 11:06:30 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:25.281 11:06:30 -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:17:25.281 11:06:30 -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:25.281 11:06:30 -- common/autotest_common.sh@10 -- # set +x 00:17:25.281 ************************************ 00:17:25.281 START TEST blockdev_raid5f 00:17:25.281 ************************************ 00:17:25.281 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:25.281 * Looking for test storage... 00:17:25.281 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:17:25.281 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1690 -- # [[ y == y ]] 00:17:25.281 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1691 -- # lcov --version 00:17:25.281 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1691 -- # awk '{print $NF}' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1691 -- # lt 1.15 2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:25.542 11:06:30 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1692 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1704 -- # export 'LCOV_OPTS= 00:17:25.542 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:25.542 --rc genhtml_branch_coverage=1 00:17:25.542 --rc genhtml_function_coverage=1 00:17:25.542 --rc genhtml_legend=1 00:17:25.542 --rc geninfo_all_blocks=1 00:17:25.542 --rc geninfo_unexecuted_blocks=1 00:17:25.542 00:17:25.542 ' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1704 -- # LCOV_OPTS=' 00:17:25.542 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:25.542 --rc genhtml_branch_coverage=1 00:17:25.542 --rc genhtml_function_coverage=1 00:17:25.542 --rc genhtml_legend=1 00:17:25.542 --rc geninfo_all_blocks=1 00:17:25.542 --rc geninfo_unexecuted_blocks=1 00:17:25.542 00:17:25.542 ' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1705 -- # export 'LCOV=lcov 00:17:25.542 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:25.542 --rc genhtml_branch_coverage=1 00:17:25.542 --rc genhtml_function_coverage=1 00:17:25.542 --rc genhtml_legend=1 00:17:25.542 --rc geninfo_all_blocks=1 00:17:25.542 --rc geninfo_unexecuted_blocks=1 00:17:25.542 00:17:25.542 ' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@1705 -- # LCOV='lcov 00:17:25.542 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:25.542 --rc genhtml_branch_coverage=1 00:17:25.542 --rc genhtml_function_coverage=1 00:17:25.542 --rc genhtml_legend=1 00:17:25.542 --rc geninfo_all_blocks=1 00:17:25.542 --rc geninfo_unexecuted_blocks=1 00:17:25.542 00:17:25.542 ' 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@669 -- # QOS_DEV_1=Malloc_0 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@670 -- # QOS_DEV_2=Null_1 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_RUN_TIME=5 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@673 -- # uname -s 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@673 -- # '[' Linux = Linux ']' 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@675 -- # PRE_RESERVED_MEM=0 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@681 -- # test_type=raid5f 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@682 -- # crypto_device= 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@683 -- # dek= 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@684 -- # env_ctx= 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@685 -- # wait_for_rpc= 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@686 -- # '[' -n '' ']' 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == bdev ]] 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == crypto_* ]] 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@692 -- # start_spdk_tgt 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=100375 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:17:25.542 11:06:30 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 100375 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@833 -- # '[' -z 100375 ']' 00:17:25.542 11:06:30 blockdev_raid5f -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:25.543 11:06:30 blockdev_raid5f -- common/autotest_common.sh@838 -- # local max_retries=100 00:17:25.543 11:06:30 blockdev_raid5f -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:25.543 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:25.543 11:06:30 blockdev_raid5f -- common/autotest_common.sh@842 -- # xtrace_disable 00:17:25.543 11:06:30 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:25.543 [2024-10-29 11:06:30.948051] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:25.543 [2024-10-29 11:06:30.948250] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100375 ] 00:17:25.802 [2024-10-29 11:06:31.116487] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:25.802 [2024-10-29 11:06:31.143035] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@866 -- # return 0 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@693 -- # case "$test_type" in 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@725 -- # setup_raid5f_conf 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.373 Malloc0 00:17:26.373 Malloc1 00:17:26.373 Malloc2 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@736 -- # rpc_cmd bdev_wait_for_examine 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@739 -- # cat 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n accel 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.373 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n bdev 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.373 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n iobuf 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@747 -- # mapfile -t bdevs 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@747 -- # rpc_cmd bdev_get_bdevs 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@747 -- # jq -r '.[] | select(.claimed == false)' 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@748 -- # mapfile -t bdevs_name 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@748 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "3b17455c-fae6-4bde-8a4d-13f49917c406"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "3b17455c-fae6-4bde-8a4d-13f49917c406",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "3b17455c-fae6-4bde-8a4d-13f49917c406",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "8f4f7f64-ca00-4e22-a957-62ff5a57d4b5",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "fe1e10d0-6272-4168-b8a6-9cd8266c8fa7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8e60afa8-4373-4987-abc0-03f302bcf14a",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@748 -- # jq -r .name 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@749 -- # bdev_list=("${bdevs_name[@]}") 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@751 -- # hello_world_bdev=raid5f 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@752 -- # trap - SIGINT SIGTERM EXIT 00:17:26.633 11:06:31 blockdev_raid5f -- bdev/blockdev.sh@753 -- # killprocess 100375 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@952 -- # '[' -z 100375 ']' 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@956 -- # kill -0 100375 00:17:26.633 11:06:31 blockdev_raid5f -- common/autotest_common.sh@957 -- # uname 00:17:26.633 11:06:32 blockdev_raid5f -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:26.633 11:06:32 blockdev_raid5f -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 100375 00:17:26.634 killing process with pid 100375 00:17:26.634 11:06:32 blockdev_raid5f -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:26.634 11:06:32 blockdev_raid5f -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:26.634 11:06:32 blockdev_raid5f -- common/autotest_common.sh@970 -- # echo 'killing process with pid 100375' 00:17:26.634 11:06:32 blockdev_raid5f -- common/autotest_common.sh@971 -- # kill 100375 00:17:26.634 11:06:32 blockdev_raid5f -- common/autotest_common.sh@976 -- # wait 100375 00:17:27.204 11:06:32 blockdev_raid5f -- bdev/blockdev.sh@757 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:27.204 11:06:32 blockdev_raid5f -- bdev/blockdev.sh@759 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:27.204 11:06:32 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 7 -le 1 ']' 00:17:27.204 11:06:32 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:27.204 11:06:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:27.204 ************************************ 00:17:27.204 START TEST bdev_hello_world 00:17:27.204 ************************************ 00:17:27.204 11:06:32 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:27.204 [2024-10-29 11:06:32.538501] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:27.204 [2024-10-29 11:06:32.538697] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100409 ] 00:17:27.464 [2024-10-29 11:06:32.711998] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:27.464 [2024-10-29 11:06:32.737933] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:27.464 [2024-10-29 11:06:32.918213] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:17:27.464 [2024-10-29 11:06:32.918376] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:17:27.464 [2024-10-29 11:06:32.918415] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:17:27.464 [2024-10-29 11:06:32.918827] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:17:27.464 [2024-10-29 11:06:32.919023] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:17:27.464 [2024-10-29 11:06:32.919088] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:17:27.464 [2024-10-29 11:06:32.919194] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:17:27.464 00:17:27.464 [2024-10-29 11:06:32.919262] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:17:27.724 00:17:27.724 ************************************ 00:17:27.724 END TEST bdev_hello_world 00:17:27.724 ************************************ 00:17:27.724 real 0m0.699s 00:17:27.724 user 0m0.361s 00:17:27.724 sys 0m0.225s 00:17:27.724 11:06:33 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:27.724 11:06:33 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:17:27.724 11:06:33 blockdev_raid5f -- bdev/blockdev.sh@760 -- # run_test bdev_bounds bdev_bounds '' 00:17:27.724 11:06:33 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:17:27.724 11:06:33 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:27.724 11:06:33 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:27.984 ************************************ 00:17:27.984 START TEST bdev_bounds 00:17:27.984 ************************************ 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1127 -- # bdev_bounds '' 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=100442 00:17:27.984 Process bdevio pid: 100442 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 100442' 00:17:27.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 100442 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@833 -- # '[' -z 100442 ']' 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@838 -- # local max_retries=100 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # xtrace_disable 00:17:27.984 11:06:33 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:27.984 [2024-10-29 11:06:33.325375] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:27.984 [2024-10-29 11:06:33.325543] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100442 ] 00:17:28.244 [2024-10-29 11:06:33.501481] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:17:28.244 [2024-10-29 11:06:33.530538] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:28.244 [2024-10-29 11:06:33.530648] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:28.244 [2024-10-29 11:06:33.530749] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:17:28.814 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:17:28.814 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@866 -- # return 0 00:17:28.814 11:06:34 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:17:28.814 I/O targets: 00:17:28.814 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:17:28.814 00:17:28.814 00:17:28.814 CUnit - A unit testing framework for C - Version 2.1-3 00:17:28.814 http://cunit.sourceforge.net/ 00:17:28.814 00:17:28.814 00:17:28.814 Suite: bdevio tests on: raid5f 00:17:28.814 Test: blockdev write read block ...passed 00:17:28.814 Test: blockdev write zeroes read block ...passed 00:17:28.814 Test: blockdev write zeroes read no split ...passed 00:17:28.815 Test: blockdev write zeroes read split ...passed 00:17:29.074 Test: blockdev write zeroes read split partial ...passed 00:17:29.074 Test: blockdev reset ...passed 00:17:29.074 Test: blockdev write read 8 blocks ...passed 00:17:29.074 Test: blockdev write read size > 128k ...passed 00:17:29.074 Test: blockdev write read invalid size ...passed 00:17:29.074 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:17:29.074 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:17:29.074 Test: blockdev write read max offset ...passed 00:17:29.075 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:17:29.075 Test: blockdev writev readv 8 blocks ...passed 00:17:29.075 Test: blockdev writev readv 30 x 1block ...passed 00:17:29.075 Test: blockdev writev readv block ...passed 00:17:29.075 Test: blockdev writev readv size > 128k ...passed 00:17:29.075 Test: blockdev writev readv size > 128k in two iovs ...passed 00:17:29.075 Test: blockdev comparev and writev ...passed 00:17:29.075 Test: blockdev nvme passthru rw ...passed 00:17:29.075 Test: blockdev nvme passthru vendor specific ...passed 00:17:29.075 Test: blockdev nvme admin passthru ...passed 00:17:29.075 Test: blockdev copy ...passed 00:17:29.075 00:17:29.075 Run Summary: Type Total Ran Passed Failed Inactive 00:17:29.075 suites 1 1 n/a 0 0 00:17:29.075 tests 23 23 23 0 0 00:17:29.075 asserts 130 130 130 0 n/a 00:17:29.075 00:17:29.075 Elapsed time = 0.309 seconds 00:17:29.075 0 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 100442 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@952 -- # '[' -z 100442 ']' 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # kill -0 100442 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@957 -- # uname 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 100442 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@970 -- # echo 'killing process with pid 100442' 00:17:29.075 killing process with pid 100442 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@971 -- # kill 100442 00:17:29.075 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@976 -- # wait 100442 00:17:29.335 11:06:34 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:17:29.335 00:17:29.335 real 0m1.449s 00:17:29.335 user 0m3.451s 00:17:29.335 sys 0m0.376s 00:17:29.335 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:29.335 11:06:34 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:29.335 ************************************ 00:17:29.335 END TEST bdev_bounds 00:17:29.335 ************************************ 00:17:29.335 11:06:34 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:29.335 11:06:34 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 5 -le 1 ']' 00:17:29.335 11:06:34 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:29.335 11:06:34 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:29.335 ************************************ 00:17:29.335 START TEST bdev_nbd 00:17:29.335 ************************************ 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1127 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100485 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100485 /var/tmp/spdk-nbd.sock 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@833 -- # '[' -z 100485 ']' 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@837 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@838 -- # local max_retries=100 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:29.336 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # xtrace_disable 00:17:29.336 11:06:34 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:29.596 [2024-10-29 11:06:34.867757] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:29.596 [2024-10-29 11:06:34.867987] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:29.596 [2024-10-29 11:06:35.047264] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:29.596 [2024-10-29 11:06:35.074686] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@862 -- # (( i == 0 )) 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@866 -- # return 0 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:17:30.537 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # local i 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # break 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:30.538 1+0 records in 00:17:30.538 1+0 records out 00:17:30.538 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000363548 s, 11.3 MB/s 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # size=4096 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # return 0 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:30.538 11:06:35 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:17:30.798 { 00:17:30.798 "nbd_device": "/dev/nbd0", 00:17:30.798 "bdev_name": "raid5f" 00:17:30.798 } 00:17:30.798 ]' 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:17:30.798 { 00:17:30.798 "nbd_device": "/dev/nbd0", 00:17:30.798 "bdev_name": "raid5f" 00:17:30.798 } 00:17:30.798 ]' 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:30.798 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.058 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:31.318 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:17:31.578 /dev/nbd0 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@870 -- # local nbd_name=nbd0 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # local i 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # (( i = 1 )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # (( i <= 20 )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@874 -- # grep -q -w nbd0 /proc/partitions 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # break 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # (( i = 1 )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # (( i <= 20 )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:31.578 1+0 records in 00:17:31.578 1+0 records out 00:17:31.578 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00046583 s, 8.8 MB/s 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # size=4096 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # '[' 4096 '!=' 0 ']' 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # return 0 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.578 11:06:36 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:31.837 { 00:17:31.837 "nbd_device": "/dev/nbd0", 00:17:31.837 "bdev_name": "raid5f" 00:17:31.837 } 00:17:31.837 ]' 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:31.837 { 00:17:31.837 "nbd_device": "/dev/nbd0", 00:17:31.837 "bdev_name": "raid5f" 00:17:31.837 } 00:17:31.837 ]' 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:31.837 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:17:31.838 256+0 records in 00:17:31.838 256+0 records out 00:17:31.838 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0143016 s, 73.3 MB/s 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:31.838 256+0 records in 00:17:31.838 256+0 records out 00:17:31.838 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0317268 s, 33.1 MB/s 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:31.838 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.098 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:17:32.358 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:17:32.618 malloc_lvol_verify 00:17:32.618 11:06:37 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:17:32.878 e4bb18f3-4d17-4712-80e4-7b457ae81feb 00:17:32.878 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:17:32.878 d016f8af-bd1b-456f-8ee4-e3478cb2383b 00:17:32.878 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:17:33.139 /dev/nbd0 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:17:33.139 mke2fs 1.47.0 (5-Feb-2023) 00:17:33.139 Discarding device blocks: 0/4096 done 00:17:33.139 Creating filesystem with 4096 1k blocks and 1024 inodes 00:17:33.139 00:17:33.139 Allocating group tables: 0/1 done 00:17:33.139 Writing inode tables: 0/1 done 00:17:33.139 Creating journal (1024 blocks): done 00:17:33.139 Writing superblocks and filesystem accounting information: 0/1 done 00:17:33.139 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:33.139 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100485 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@952 -- # '[' -z 100485 ']' 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # kill -0 100485 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@957 -- # uname 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@957 -- # '[' Linux = Linux ']' 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # ps --no-headers -o comm= 100485 00:17:33.400 killing process with pid 100485 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # process_name=reactor_0 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@962 -- # '[' reactor_0 = sudo ']' 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@970 -- # echo 'killing process with pid 100485' 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@971 -- # kill 100485 00:17:33.400 11:06:38 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@976 -- # wait 100485 00:17:33.660 11:06:39 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:17:33.660 00:17:33.660 real 0m4.336s 00:17:33.660 user 0m6.211s 00:17:33.660 sys 0m1.338s 00:17:33.660 11:06:39 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:33.660 ************************************ 00:17:33.660 END TEST bdev_nbd 00:17:33.660 ************************************ 00:17:33.660 11:06:39 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:33.660 11:06:39 blockdev_raid5f -- bdev/blockdev.sh@762 -- # [[ y == y ]] 00:17:33.660 11:06:39 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = nvme ']' 00:17:33.660 11:06:39 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = gpt ']' 00:17:33.660 11:06:39 blockdev_raid5f -- bdev/blockdev.sh@767 -- # run_test bdev_fio fio_test_suite '' 00:17:33.660 11:06:39 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 3 -le 1 ']' 00:17:33.660 11:06:39 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:33.660 11:06:39 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:33.921 ************************************ 00:17:33.921 START TEST bdev_fio 00:17:33.921 ************************************ 00:17:33.921 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1127 -- # fio_test_suite '' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local workload=verify 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local bdev_type=AIO 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local env_context= 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local fio_dir=/usr/src/fio 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1293 -- # '[' -z verify ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1297 -- # '[' -n '' ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # cat 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1315 -- # '[' verify == verify ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1316 -- # cat 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1325 -- # '[' AIO == AIO ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1326 -- # /usr/src/fio/fio --version 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1326 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # echo serialize_overlap=1 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1103 -- # '[' 11 -le 1 ']' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:33.921 ************************************ 00:17:33.921 START TEST bdev_fio_rw_verify 00:17:33.921 ************************************ 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1127 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1358 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local fio_dir=/usr/src/fio 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local sanitizers 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1342 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # shift 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # local asan_lib= 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # for sanitizer in "${sanitizers[@]}" 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # grep libasan 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # awk '{print $3}' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # asan_lib=/usr/lib64/libasan.so.8 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # break 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1354 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:17:33.921 11:06:39 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1354 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:34.182 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:17:34.182 fio-3.35 00:17:34.182 Starting 1 thread 00:17:46.473 00:17:46.473 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100672: Tue Oct 29 11:06:50 2024 00:17:46.473 read: IOPS=12.8k, BW=50.0MiB/s (52.4MB/s)(500MiB/10001msec) 00:17:46.473 slat (usec): min=16, max=248, avg=18.27, stdev= 2.20 00:17:46.473 clat (usec): min=11, max=479, avg=125.17, stdev=43.48 00:17:46.473 lat (usec): min=29, max=501, avg=143.44, stdev=43.66 00:17:46.473 clat percentiles (usec): 00:17:46.473 | 50.000th=[ 128], 99.000th=[ 206], 99.900th=[ 233], 99.990th=[ 412], 00:17:46.473 | 99.999th=[ 474] 00:17:46.473 write: IOPS=13.4k, BW=52.4MiB/s (54.9MB/s)(518MiB/9879msec); 0 zone resets 00:17:46.473 slat (usec): min=7, max=305, avg=15.91, stdev= 4.15 00:17:46.473 clat (usec): min=56, max=1741, avg=287.16, stdev=49.51 00:17:46.473 lat (usec): min=71, max=1990, avg=303.07, stdev=51.22 00:17:46.473 clat percentiles (usec): 00:17:46.473 | 50.000th=[ 293], 99.000th=[ 367], 99.900th=[ 889], 99.990th=[ 1483], 00:17:46.473 | 99.999th=[ 1647] 00:17:46.473 bw ( KiB/s): min=48864, max=56072, per=98.98%, avg=53112.89, stdev=1886.29, samples=19 00:17:46.473 iops : min=12212, max=14022, avg=13278.21, stdev=472.77, samples=19 00:17:46.473 lat (usec) : 20=0.01%, 50=0.01%, 100=17.16%, 250=41.33%, 500=41.32% 00:17:46.473 lat (usec) : 750=0.10%, 1000=0.04% 00:17:46.473 lat (msec) : 2=0.03% 00:17:46.473 cpu : usr=98.93%, sys=0.37%, ctx=25, majf=0, minf=13523 00:17:46.473 IO depths : 1=7.6%, 2=19.9%, 4=55.1%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:46.473 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:46.473 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:46.473 issued rwts: total=128020,132521,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:46.473 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:46.473 00:17:46.473 Run status group 0 (all jobs): 00:17:46.473 READ: bw=50.0MiB/s (52.4MB/s), 50.0MiB/s-50.0MiB/s (52.4MB/s-52.4MB/s), io=500MiB (524MB), run=10001-10001msec 00:17:46.473 WRITE: bw=52.4MiB/s (54.9MB/s), 52.4MiB/s-52.4MiB/s (54.9MB/s-54.9MB/s), io=518MiB (543MB), run=9879-9879msec 00:17:46.473 ----------------------------------------------------- 00:17:46.473 Suppressions used: 00:17:46.473 count bytes template 00:17:46.473 1 7 /usr/src/fio/parse.c 00:17:46.473 522 50112 /usr/src/fio/iolog.c 00:17:46.473 1 8 libtcmalloc_minimal.so 00:17:46.473 1 904 libcrypto.so 00:17:46.473 ----------------------------------------------------- 00:17:46.473 00:17:46.473 00:17:46.473 real 0m11.221s 00:17:46.473 user 0m11.446s 00:17:46.473 sys 0m0.539s 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:46.473 ************************************ 00:17:46.473 END TEST bdev_fio_rw_verify 00:17:46.473 ************************************ 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local workload=trim 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local bdev_type= 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local env_context= 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local fio_dir=/usr/src/fio 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1293 -- # '[' -z trim ']' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1297 -- # '[' -n '' ']' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # cat 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1315 -- # '[' trim == verify ']' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1330 -- # '[' trim == trim ']' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1331 -- # echo rw=trimwrite 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "3b17455c-fae6-4bde-8a4d-13f49917c406"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "3b17455c-fae6-4bde-8a4d-13f49917c406",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "3b17455c-fae6-4bde-8a4d-13f49917c406",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "8f4f7f64-ca00-4e22-a957-62ff5a57d4b5",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "fe1e10d0-6272-4168-b8a6-9cd8266c8fa7",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8e60afa8-4373-4987-abc0-03f302bcf14a",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:46.473 /home/vagrant/spdk_repo/spdk 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:46.473 00:17:46.473 real 0m11.509s 00:17:46.473 user 0m11.555s 00:17:46.473 sys 0m0.683s 00:17:46.473 ************************************ 00:17:46.473 END TEST bdev_fio 00:17:46.473 ************************************ 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:46.473 11:06:50 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:46.474 11:06:50 blockdev_raid5f -- bdev/blockdev.sh@774 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:46.474 11:06:50 blockdev_raid5f -- bdev/blockdev.sh@776 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:46.474 11:06:50 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 16 -le 1 ']' 00:17:46.474 11:06:50 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:46.474 11:06:50 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:46.474 ************************************ 00:17:46.474 START TEST bdev_verify 00:17:46.474 ************************************ 00:17:46.474 11:06:50 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:46.474 [2024-10-29 11:06:50.835539] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:46.474 [2024-10-29 11:06:50.835724] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100834 ] 00:17:46.474 [2024-10-29 11:06:51.011203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:46.474 [2024-10-29 11:06:51.038718] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:46.474 [2024-10-29 11:06:51.038768] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:46.474 Running I/O for 5 seconds... 00:17:47.984 11070.00 IOPS, 43.24 MiB/s [2024-10-29T11:06:54.422Z] 11185.00 IOPS, 43.69 MiB/s [2024-10-29T11:06:55.362Z] 11220.67 IOPS, 43.83 MiB/s [2024-10-29T11:06:56.303Z] 11212.25 IOPS, 43.80 MiB/s [2024-10-29T11:06:56.303Z] 11208.80 IOPS, 43.78 MiB/s 00:17:50.805 Latency(us) 00:17:50.805 [2024-10-29T11:06:56.303Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:50.805 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:50.805 Verification LBA range: start 0x0 length 0x2000 00:17:50.805 raid5f : 5.02 4501.93 17.59 0.00 0.00 42612.73 227.16 30220.97 00:17:50.805 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:50.805 Verification LBA range: start 0x2000 length 0x2000 00:17:50.805 raid5f : 5.02 6710.26 26.21 0.00 0.00 28680.25 266.51 21749.94 00:17:50.805 [2024-10-29T11:06:56.303Z] =================================================================================================================== 00:17:50.805 [2024-10-29T11:06:56.303Z] Total : 11212.19 43.80 0.00 0.00 34276.26 227.16 30220.97 00:17:51.065 00:17:51.065 real 0m5.739s 00:17:51.065 user 0m10.674s 00:17:51.065 sys 0m0.250s 00:17:51.065 11:06:56 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:51.065 11:06:56 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:51.065 ************************************ 00:17:51.065 END TEST bdev_verify 00:17:51.065 ************************************ 00:17:51.065 11:06:56 blockdev_raid5f -- bdev/blockdev.sh@777 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:51.065 11:06:56 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 16 -le 1 ']' 00:17:51.065 11:06:56 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:51.065 11:06:56 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:51.065 ************************************ 00:17:51.065 START TEST bdev_verify_big_io 00:17:51.065 ************************************ 00:17:51.065 11:06:56 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:51.326 [2024-10-29 11:06:56.654185] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:51.326 [2024-10-29 11:06:56.654402] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100917 ] 00:17:51.587 [2024-10-29 11:06:56.832395] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:51.587 [2024-10-29 11:06:56.862229] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:51.587 [2024-10-29 11:06:56.862315] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:51.587 Running I/O for 5 seconds... 00:17:53.911 633.00 IOPS, 39.56 MiB/s [2024-10-29T11:07:00.349Z] 761.00 IOPS, 47.56 MiB/s [2024-10-29T11:07:01.288Z] 803.00 IOPS, 50.19 MiB/s [2024-10-29T11:07:02.228Z] 793.25 IOPS, 49.58 MiB/s [2024-10-29T11:07:02.488Z] 812.40 IOPS, 50.77 MiB/s 00:17:56.990 Latency(us) 00:17:56.990 [2024-10-29T11:07:02.488Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:56.990 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:56.990 Verification LBA range: start 0x0 length 0x200 00:17:56.990 raid5f : 5.27 361.45 22.59 0.00 0.00 8763598.12 206.59 371809.48 00:17:56.990 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:56.990 Verification LBA range: start 0x200 length 0x200 00:17:56.990 raid5f : 5.22 462.31 28.89 0.00 0.00 6933873.49 152.93 298546.53 00:17:56.990 [2024-10-29T11:07:02.488Z] =================================================================================================================== 00:17:56.990 [2024-10-29T11:07:02.488Z] Total : 823.76 51.48 0.00 0.00 7741104.94 152.93 371809.48 00:17:57.250 00:17:57.250 real 0m5.988s 00:17:57.250 user 0m11.163s 00:17:57.250 sys 0m0.243s 00:17:57.250 ************************************ 00:17:57.250 END TEST bdev_verify_big_io 00:17:57.250 ************************************ 00:17:57.250 11:07:02 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:57.250 11:07:02 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:57.250 11:07:02 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:57.250 11:07:02 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 13 -le 1 ']' 00:17:57.250 11:07:02 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:57.250 11:07:02 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:57.250 ************************************ 00:17:57.250 START TEST bdev_write_zeroes 00:17:57.250 ************************************ 00:17:57.250 11:07:02 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:57.250 [2024-10-29 11:07:02.715912] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:57.250 [2024-10-29 11:07:02.716093] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100993 ] 00:17:57.509 [2024-10-29 11:07:02.891801] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:57.509 [2024-10-29 11:07:02.919682] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:57.769 Running I/O for 1 seconds... 00:17:58.709 30471.00 IOPS, 119.03 MiB/s 00:17:58.709 Latency(us) 00:17:58.709 [2024-10-29T11:07:04.207Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:58.709 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:58.709 raid5f : 1.01 30435.68 118.89 0.00 0.00 4194.60 1330.75 5695.05 00:17:58.709 [2024-10-29T11:07:04.207Z] =================================================================================================================== 00:17:58.709 [2024-10-29T11:07:04.207Z] Total : 30435.68 118.89 0.00 0.00 4194.60 1330.75 5695.05 00:17:58.970 00:17:58.970 real 0m1.705s 00:17:58.970 user 0m1.370s 00:17:58.970 sys 0m0.222s 00:17:58.970 11:07:04 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:58.970 ************************************ 00:17:58.970 END TEST bdev_write_zeroes 00:17:58.970 ************************************ 00:17:58.970 11:07:04 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:58.970 11:07:04 blockdev_raid5f -- bdev/blockdev.sh@781 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:58.970 11:07:04 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 13 -le 1 ']' 00:17:58.970 11:07:04 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:58.970 11:07:04 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:58.970 ************************************ 00:17:58.970 START TEST bdev_json_nonenclosed 00:17:58.970 ************************************ 00:17:58.970 11:07:04 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:59.230 [2024-10-29 11:07:04.497868] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:59.230 [2024-10-29 11:07:04.498056] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101035 ] 00:17:59.230 [2024-10-29 11:07:04.671453] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:59.230 [2024-10-29 11:07:04.699514] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:59.230 [2024-10-29 11:07:04.699636] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:59.231 [2024-10-29 11:07:04.699657] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:59.231 [2024-10-29 11:07:04.699675] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:59.491 ************************************ 00:17:59.491 END TEST bdev_json_nonenclosed 00:17:59.491 ************************************ 00:17:59.491 00:17:59.491 real 0m0.381s 00:17:59.491 user 0m0.149s 00:17:59.491 sys 0m0.127s 00:17:59.491 11:07:04 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:59.491 11:07:04 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:59.491 11:07:04 blockdev_raid5f -- bdev/blockdev.sh@784 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:59.491 11:07:04 blockdev_raid5f -- common/autotest_common.sh@1103 -- # '[' 13 -le 1 ']' 00:17:59.491 11:07:04 blockdev_raid5f -- common/autotest_common.sh@1109 -- # xtrace_disable 00:17:59.491 11:07:04 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:59.491 ************************************ 00:17:59.491 START TEST bdev_json_nonarray 00:17:59.491 ************************************ 00:17:59.491 11:07:04 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1127 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:59.491 [2024-10-29 11:07:04.945825] Starting SPDK v25.01-pre git sha1 12fc2abf1 / DPDK 23.11.0 initialization... 00:17:59.491 [2024-10-29 11:07:04.945940] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid101061 ] 00:17:59.752 [2024-10-29 11:07:05.116257] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:59.752 [2024-10-29 11:07:05.146299] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:59.752 [2024-10-29 11:07:05.146434] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:59.752 [2024-10-29 11:07:05.146456] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:59.752 [2024-10-29 11:07:05.146469] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:59.752 00:17:59.752 real 0m0.374s 00:17:59.752 user 0m0.149s 00:17:59.752 sys 0m0.121s 00:17:59.752 ************************************ 00:17:59.752 END TEST bdev_json_nonarray 00:17:59.752 ************************************ 00:17:59.752 11:07:05 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1128 -- # xtrace_disable 00:17:59.752 11:07:05 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@786 -- # [[ raid5f == bdev ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@793 -- # [[ raid5f == gpt ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@797 -- # [[ raid5f == crypto_sw ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@809 -- # trap - SIGINT SIGTERM EXIT 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@810 -- # cleanup 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:18:00.013 11:07:05 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:18:00.013 ************************************ 00:18:00.013 END TEST blockdev_raid5f 00:18:00.013 ************************************ 00:18:00.013 00:18:00.013 real 0m34.708s 00:18:00.013 user 0m46.993s 00:18:00.013 sys 0m4.677s 00:18:00.013 11:07:05 blockdev_raid5f -- common/autotest_common.sh@1128 -- # xtrace_disable 00:18:00.013 11:07:05 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:00.013 11:07:05 -- spdk/autotest.sh@194 -- # uname -s 00:18:00.013 11:07:05 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@256 -- # timing_exit lib 00:18:00.013 11:07:05 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:00.013 11:07:05 -- common/autotest_common.sh@10 -- # set +x 00:18:00.013 11:07:05 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@272 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:18:00.013 11:07:05 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:18:00.013 11:07:05 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:18:00.013 11:07:05 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:18:00.013 11:07:05 -- common/autotest_common.sh@724 -- # xtrace_disable 00:18:00.013 11:07:05 -- common/autotest_common.sh@10 -- # set +x 00:18:00.013 11:07:05 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:18:00.013 11:07:05 -- common/autotest_common.sh@1394 -- # local autotest_es=0 00:18:00.013 11:07:05 -- common/autotest_common.sh@1395 -- # xtrace_disable 00:18:00.013 11:07:05 -- common/autotest_common.sh@10 -- # set +x 00:18:02.557 INFO: APP EXITING 00:18:02.557 INFO: killing all VMs 00:18:02.557 INFO: killing vhost app 00:18:02.557 INFO: EXIT DONE 00:18:02.817 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:02.817 Waiting for block devices as requested 00:18:03.077 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:18:03.077 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:18:04.019 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:04.019 Cleaning 00:18:04.019 Removing: /var/run/dpdk/spdk0/config 00:18:04.019 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:18:04.019 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:18:04.019 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:18:04.019 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:18:04.019 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:18:04.019 Removing: /var/run/dpdk/spdk0/hugepage_info 00:18:04.019 Removing: /dev/shm/spdk_tgt_trace.pid69264 00:18:04.019 Removing: /var/run/dpdk/spdk0 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100117 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100375 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100409 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100442 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100667 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100834 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100917 00:18:04.019 Removing: /var/run/dpdk/spdk_pid100993 00:18:04.278 Removing: /var/run/dpdk/spdk_pid101035 00:18:04.278 Removing: /var/run/dpdk/spdk_pid101061 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69096 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69264 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69472 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69554 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69583 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69694 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69711 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69895 00:18:04.278 Removing: /var/run/dpdk/spdk_pid69974 00:18:04.278 Removing: /var/run/dpdk/spdk_pid70059 00:18:04.278 Removing: /var/run/dpdk/spdk_pid70148 00:18:04.278 Removing: /var/run/dpdk/spdk_pid70234 00:18:04.278 Removing: /var/run/dpdk/spdk_pid70268 00:18:04.278 Removing: /var/run/dpdk/spdk_pid70310 00:18:04.279 Removing: /var/run/dpdk/spdk_pid70375 00:18:04.279 Removing: /var/run/dpdk/spdk_pid70476 00:18:04.279 Removing: /var/run/dpdk/spdk_pid70903 00:18:04.279 Removing: /var/run/dpdk/spdk_pid70956 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71008 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71019 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71094 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71110 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71172 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71184 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71237 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71255 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71297 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71315 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71444 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71486 00:18:04.279 Removing: /var/run/dpdk/spdk_pid71564 00:18:04.279 Removing: /var/run/dpdk/spdk_pid72737 00:18:04.279 Removing: /var/run/dpdk/spdk_pid72938 00:18:04.279 Removing: /var/run/dpdk/spdk_pid73067 00:18:04.279 Removing: /var/run/dpdk/spdk_pid73666 00:18:04.279 Removing: /var/run/dpdk/spdk_pid73867 00:18:04.279 Removing: /var/run/dpdk/spdk_pid73996 00:18:04.279 Removing: /var/run/dpdk/spdk_pid74605 00:18:04.279 Removing: /var/run/dpdk/spdk_pid74914 00:18:04.279 Removing: /var/run/dpdk/spdk_pid75043 00:18:04.279 Removing: /var/run/dpdk/spdk_pid76384 00:18:04.279 Removing: /var/run/dpdk/spdk_pid76625 00:18:04.279 Removing: /var/run/dpdk/spdk_pid76755 00:18:04.279 Removing: /var/run/dpdk/spdk_pid78096 00:18:04.279 Removing: /var/run/dpdk/spdk_pid78338 00:18:04.279 Removing: /var/run/dpdk/spdk_pid78467 00:18:04.279 Removing: /var/run/dpdk/spdk_pid79808 00:18:04.279 Removing: /var/run/dpdk/spdk_pid80237 00:18:04.279 Removing: /var/run/dpdk/spdk_pid80366 00:18:04.279 Removing: /var/run/dpdk/spdk_pid81807 00:18:04.279 Removing: /var/run/dpdk/spdk_pid82061 00:18:04.538 Removing: /var/run/dpdk/spdk_pid82195 00:18:04.538 Removing: /var/run/dpdk/spdk_pid83625 00:18:04.538 Removing: /var/run/dpdk/spdk_pid83879 00:18:04.538 Removing: /var/run/dpdk/spdk_pid84008 00:18:04.538 Removing: /var/run/dpdk/spdk_pid85438 00:18:04.538 Removing: /var/run/dpdk/spdk_pid85914 00:18:04.538 Removing: /var/run/dpdk/spdk_pid86043 00:18:04.538 Removing: /var/run/dpdk/spdk_pid86176 00:18:04.538 Removing: /var/run/dpdk/spdk_pid86572 00:18:04.538 Removing: /var/run/dpdk/spdk_pid87280 00:18:04.538 Removing: /var/run/dpdk/spdk_pid87643 00:18:04.538 Removing: /var/run/dpdk/spdk_pid88316 00:18:04.538 Removing: /var/run/dpdk/spdk_pid88737 00:18:04.538 Removing: /var/run/dpdk/spdk_pid89468 00:18:04.538 Removing: /var/run/dpdk/spdk_pid89861 00:18:04.538 Removing: /var/run/dpdk/spdk_pid91769 00:18:04.538 Removing: /var/run/dpdk/spdk_pid92191 00:18:04.538 Removing: /var/run/dpdk/spdk_pid92614 00:18:04.538 Removing: /var/run/dpdk/spdk_pid94647 00:18:04.538 Removing: /var/run/dpdk/spdk_pid95121 00:18:04.538 Removing: /var/run/dpdk/spdk_pid95621 00:18:04.538 Removing: /var/run/dpdk/spdk_pid96663 00:18:04.538 Removing: /var/run/dpdk/spdk_pid96980 00:18:04.538 Removing: /var/run/dpdk/spdk_pid97896 00:18:04.538 Removing: /var/run/dpdk/spdk_pid98210 00:18:04.538 Removing: /var/run/dpdk/spdk_pid99131 00:18:04.538 Removing: /var/run/dpdk/spdk_pid99445 00:18:04.538 Clean 00:18:04.538 11:07:09 -- common/autotest_common.sh@1451 -- # return 0 00:18:04.538 11:07:09 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:18:04.538 11:07:09 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:04.538 11:07:09 -- common/autotest_common.sh@10 -- # set +x 00:18:04.798 11:07:10 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:18:04.798 11:07:10 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:04.798 11:07:10 -- common/autotest_common.sh@10 -- # set +x 00:18:04.798 11:07:10 -- spdk/autotest.sh@388 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:04.798 11:07:10 -- spdk/autotest.sh@390 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:18:04.798 11:07:10 -- spdk/autotest.sh@390 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:18:04.798 11:07:10 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:18:04.798 11:07:10 -- spdk/autotest.sh@394 -- # hostname 00:18:04.798 11:07:10 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:18:05.058 geninfo: WARNING: invalid characters removed from testname! 00:18:31.627 11:07:34 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:31.627 11:07:36 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:33.545 11:07:38 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:35.455 11:07:40 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:37.432 11:07:42 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:39.973 11:07:44 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:41.881 11:07:47 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:41.881 11:07:47 -- spdk/autorun.sh@1 -- $ timing_finish 00:18:41.881 11:07:47 -- common/autotest_common.sh@736 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:18:41.881 11:07:47 -- common/autotest_common.sh@738 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:41.881 11:07:47 -- common/autotest_common.sh@739 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:41.881 11:07:47 -- common/autotest_common.sh@742 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:41.881 + [[ -n 6152 ]] 00:18:41.881 + sudo kill 6152 00:18:41.890 [Pipeline] } 00:18:41.902 [Pipeline] // timeout 00:18:41.905 [Pipeline] } 00:18:41.976 [Pipeline] // stage 00:18:41.980 [Pipeline] } 00:18:41.993 [Pipeline] // catchError 00:18:42.000 [Pipeline] stage 00:18:42.001 [Pipeline] { (Stop VM) 00:18:42.012 [Pipeline] sh 00:18:42.292 + vagrant halt 00:18:44.829 ==> default: Halting domain... 00:18:52.980 [Pipeline] sh 00:18:53.269 + vagrant destroy -f 00:18:55.807 ==> default: Removing domain... 00:18:55.819 [Pipeline] sh 00:18:56.102 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:56.113 [Pipeline] } 00:18:56.127 [Pipeline] // stage 00:18:56.133 [Pipeline] } 00:18:56.147 [Pipeline] // dir 00:18:56.152 [Pipeline] } 00:18:56.167 [Pipeline] // wrap 00:18:56.173 [Pipeline] } 00:18:56.183 [Pipeline] // catchError 00:18:56.191 [Pipeline] stage 00:18:56.194 [Pipeline] { (Epilogue) 00:18:56.207 [Pipeline] sh 00:18:56.493 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:19:00.730 [Pipeline] catchError 00:19:00.733 [Pipeline] { 00:19:00.746 [Pipeline] sh 00:19:01.033 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:19:01.033 Artifacts sizes are good 00:19:01.045 [Pipeline] } 00:19:01.060 [Pipeline] // catchError 00:19:01.070 [Pipeline] archiveArtifacts 00:19:01.078 Archiving artifacts 00:19:01.202 [Pipeline] cleanWs 00:19:01.219 [WS-CLEANUP] Deleting project workspace... 00:19:01.219 [WS-CLEANUP] Deferred wipeout is used... 00:19:01.228 [WS-CLEANUP] done 00:19:01.230 [Pipeline] } 00:19:01.245 [Pipeline] // stage 00:19:01.252 [Pipeline] } 00:19:01.269 [Pipeline] // node 00:19:01.275 [Pipeline] End of Pipeline 00:19:01.332 Finished: SUCCESS